Wang, Yimin
Analysis of Higher-Order Ising Hamiltonians
Cen, Yunuo, Zhang, Zhiwei, Wang, Zixuan, Wang, Yimin, Fong, Xuanyao
It is challenging to scale Ising machines for industrial-level problems due to algorithm or hardware limitations. Although higher-order Ising models provide a more compact encoding, they are, however, hard to physically implement. This work proposes a theoretical framework of a higher-order Ising simulator, IsingSim. The Ising spins and gradients in IsingSim are decoupled and self-customizable. We significantly accelerate the simulation speed via a bidirectional approach for differentiating the hyperedge functions. Our proof-of-concept implementation verifies the theoretical framework by simulating the Ising spins with exact and approximate gradients. Experiment results show that our novel framework can be a useful tool for providing design guidelines for higher-order Ising machines.
Instruction Tuning Vs. In-Context Learning: Revisiting Large Language Models in Few-Shot Computational Social Science
Wang, Taihang, Xu, Xiaoman, Wang, Yimin, Jiang, Ye
Real-world applications of large language models (LLMs) in computational social science (CSS) tasks primarily depend on the effectiveness of instruction tuning (IT) or in-context learning (ICL). While IT has shown highly effective at fine-tuning LLMs for various tasks, ICL offers a rapid alternative for task adaptation by learning from examples without explicit gradient updates. In this paper, we evaluate the classification performance of LLMs using IT versus ICL in few-shot CSS tasks. The experimental results indicate that ICL consistently outperforms IT in most CSS tasks. Additionally, we investigate the relationship between the increasing number of training samples and LLM performance. Our findings show that simply increasing the number of samples without considering their quality does not consistently enhance the performance of LLMs with either ICL or IT and can sometimes even result in a performance decline. Finally, we compare three prompting strategies, demonstrating that ICL is more effective than zero-shot and Chain-of-Thought (CoT). Our research highlights the significant advantages of ICL in handling CSS tasks in few-shot settings and emphasizes the importance of optimizing sample quality and prompting strategies to improve LLM classification performance. The code will be made available.
Large Visual-Language Models Are Also Good Classifiers: A Study of In-Context Multimodal Fake News Detection
Jiang, Ye, Wang, Yimin
Large visual-language models (LVLMs) exhibit exceptional performance in visual-language reasoning across diverse cross-modal benchmarks. Despite these advances, recent research indicates that Large Language Models (LLMs), like GPT-3.5-turbo, underachieve compared to well-trained smaller models, such as BERT, in Fake News Detection (FND), prompting inquiries into LVLMs' efficacy in FND tasks. Although performance could improve through fine-tuning LVLMs, the substantial parameters and requisite pre-trained weights render it a resource-heavy endeavor for FND applications. This paper initially assesses the FND capabilities of two notable LVLMs, CogVLM and GPT4V, in comparison to a smaller yet adeptly trained CLIP model in a zero-shot context. The findings demonstrate that LVLMs can attain performance competitive with that of the smaller model. Next, we integrate standard in-context learning (ICL) with LVLMs, noting improvements in FND performance, though limited in scope and consistency. To address this, we introduce the \textbf{I}n-context \textbf{M}ultimodal \textbf{F}ake \textbf{N}ews \textbf{D}etection (IMFND) framework, enriching in-context examples and test inputs with predictions and corresponding probabilities from a well-trained smaller model. This strategic integration directs the LVLMs' focus towards news segments associated with higher probabilities, thereby improving their analytical accuracy. The experimental results suggest that the IMFND framework significantly boosts the FND efficiency of LVLMs, achieving enhanced accuracy over the standard ICL approach across three publicly available FND datasets.
Cross-Modal Augmentation for Few-Shot Multimodal Fake News Detection
Jiang, Ye, Wang, Taihang, Xu, Xiaoman, Wang, Yimin, Song, Xingyi, Maynard, Diana
The nascent topic of fake news requires automatic detection methods to quickly learn from limited annotated samples. Therefore, the capacity to rapidly acquire proficiency in a new task with limited guidance, also known as few-shot learning, is critical for detecting fake news in its early stages. Existing approaches either involve fine-tuning pre-trained language models which come with a large number of parameters, or training a complex neural network from scratch with large-scale annotated datasets. This paper presents a multimodal fake news detection model which augments multimodal features using unimodal features. For this purpose, we introduce Cross-Modal Augmentation (CMA), a simple approach for enhancing few-shot multimodal fake news detection by transforming n-shot classification into a more robust (n $\times$ z)-shot problem, where z represents the number of supplementary features. The proposed CMA achieves SOTA results over three benchmark datasets, utilizing a surprisingly simple linear probing method to classify multimodal fake news with only a few training samples. Furthermore, our method is significantly more lightweight than prior approaches, particularly in terms of the number of trainable parameters and epoch times. The code is available here: \url{https://github.com/zgjiangtoby/FND_fewshot}
A Synchronized Layer-by-layer Growing Approach for Plausible Neuronal Morphology Generation
Yang, Nianzu, Zeng, Kaipeng, Lu, Haotian, Wu, Yexin, Yuan, Zexin, Jiang, Shengdian, Wu, Jiaxiang, Wang, Yimin, Yan, Junchi
Neuronal morphology is essential for studying brain functioning and understanding neurodegenerative disorders. As the acquiring of real-world morphology data is expensive, computational approaches especially learning-based ones e.g. MorphVAE for morphology generation were recently studied, which are often conducted in a way of randomly augmenting a given authentic morphology to achieve plausibility. Under such a setting, this paper proposes \textbf{MorphGrower} which aims to generate more plausible morphology samples by mimicking the natural growth mechanism instead of a one-shot treatment as done in MorphVAE. Specifically, MorphGrower generates morphologies layer by layer synchronously and chooses a pair of sibling branches as the basic generation block, and the generation of each layer is conditioned on the morphological structure of previous layers and then generate morphologies via a conditional variational autoencoder with spherical latent space. Extensive experimental results on four real-world datasets demonstrate that MorphGrower outperforms MorphVAE by a notable margin. Our code will be publicly available to facilitate future research.
Similarity-Aware Multimodal Prompt Learning for Fake News Detection
Jiang, Ye, Yu, Xiaomin, Wang, Yimin, Xu, Xiaoman, Song, Xingyi, Maynard, Diana
The standard paradigm for fake news detection mainly utilizes text information to model the truthfulness of news. However, the discourse of online fake news is typically subtle and it requires expert knowledge to use textual information to debunk fake news. Recently, studies focusing on multimodal fake news detection have outperformed text-only methods. Recent approaches utilizing the pre-trained model to extract unimodal features, or fine-tuning the pre-trained model directly, have become a new paradigm for detecting fake news. Again, this paradigm either requires a large number of training instances, or updates the entire set of pre-trained model parameters, making real-world fake news detection impractical. Furthermore, traditional multimodal methods fuse the cross-modal features directly without considering that the uncorrelated semantic representation might inject noise into the multimodal features. This paper proposes a Similarity-Aware Multimodal Prompt Learning (SAMPLE) framework. First, we incorporate prompt learning into multimodal fake news detection. Prompt learning, which only tunes prompts with a frozen language model, can reduce memory usage significantly and achieve comparable performances, compared with fine-tuning. We analyse three prompt templates with a soft verbalizer to detect fake news. In addition, we introduce the similarity-aware fusing method to adaptively fuse the intensity of multimodal representation and mitigate the noise injection via uncorrelated cross-modal features. For evaluation, SAMPLE surpasses the F1 and the accuracies of previous works on two benchmark multimodal datasets, demonstrating the effectiveness of the proposed method in detecting fake news. In addition, SAMPLE also is superior to other approaches regardless of few-shot and data-rich settings.
AAAI-07 Workshop Reports
Anand, Sarabjot Singh, Bahls, Daniel, Burghart, Catherina R., Burstein, Mark, Chen, Huajun, Collins, John, Dietterich, Tom, Doyle, Jon, Drummond, Chris, Elazmeh, William, Geib, Christopher, Goldsmith, Judy, Guesgen, Hans W., Hendler, Jim, Jannach, Dietmar, Japkowicz, Nathalie, Junker, Ulrich, Kaminka, Gal A., Kobsa, Alfred, Lang, Jerome, Leake, David B., Lewis, Lundy, Ligozat, Gerard, Macskassy, Sofus, McDermott, Drew, Metzler, Ted, Mobasher, Bamshad, Nambiar, Ullas, Nie, Zaiqing, Orsvarn, Klas, O'Sullivan, Barry, Pynadath, David, Renz, Jochen, Rodriguez, Rita V., Roth-Berghofer, Thomas, Schulz, Stefan, Studer, Rudi, Wang, Yimin, Wellman, Michael
The AAAI-07 workshop program was held Sunday and Monday, July 22-23, in Vancouver, British Columbia, Canada. The program included the following thirteen workshops: (1) Acquiring Planning Knowledge via Demonstration; (2) Configuration; (3) Evaluating Architectures for Intelligence; (4) Evaluation Methods for Machine Learning; (5) Explanation-Aware Computing; (6) Human Implications of Human-Robot Interaction; (7) Intelligent Techniques for Web Personalization; (8) Plan, Activity, and Intent Recognition; (9) Preference Handling for Artificial Intelligence; (10) Semantic e-Science; (11) Spatial and Temporal Reasoning; (12) Trading Agent Design and Analysis; and (13) Information Integration on the Web.
AAAI-07 Workshop Reports
Anand, Sarabjot Singh, Bahls, Daniel, Burghart, Catherina R., Burstein, Mark, Chen, Huajun, Collins, John, Dietterich, Tom, Doyle, Jon, Drummond, Chris, Elazmeh, William, Geib, Christopher, Goldsmith, Judy, Guesgen, Hans W., Hendler, Jim, Jannach, Dietmar, Japkowicz, Nathalie, Junker, Ulrich, Kaminka, Gal A., Kobsa, Alfred, Lang, Jerome, Leake, David B., Lewis, Lundy, Ligozat, Gerard, Macskassy, Sofus, McDermott, Drew, Metzler, Ted, Mobasher, Bamshad, Nambiar, Ullas, Nie, Zaiqing, Orsvarn, Klas, O', Sullivan, Barry, Pynadath, David, Renz, Jochen, Rodriguez, Rita V., Roth-Berghofer, Thomas, Schulz, Stefan, Studer, Rudi, Wang, Yimin, Wellman, Michael
The AAAI-07 workshop program was held Sunday and Monday, July 22-23, in Vancouver, British Columbia, Canada. The program included the following thirteen workshops: (1) Acquiring Planning Knowledge via Demonstration; (2) Configuration; (3) Evaluating Architectures for Intelligence; (4) Evaluation Methods for Machine Learning; (5) Explanation-Aware Computing; (6) Human Implications of Human-Robot Interaction; (7) Intelligent Techniques for Web Personalization; (8) Plan, Activity, and Intent Recognition; (9) Preference Handling for Artificial Intelligence; (10) Semantic e-Science; (11) Spatial and Temporal Reasoning; (12) Trading Agent Design and Analysis; and (13) Information Integration on the Web.