Du, Zhicheng
SRAP-Agent: Simulating and Optimizing Scarce Resource Allocation Policy with LLM-based Agent
Ji, Jiarui, Li, Yang, Liu, Hongtao, Du, Zhicheng, Wei, Zhewei, Shen, Weiran, Qi, Qi, Lin, Yankai
Public scarce resource allocation plays a crucial role in economics as it directly influences the efficiency and equity in society. Traditional studies including theoretical model-based, empirical study-based and simulation-based methods encounter limitations due to the idealized assumption of complete information and individual rationality, as well as constraints posed by limited available data. In this work, we propose an innovative framework, SRAP-Agent (Simulating and Optimizing Scarce Resource Allocation Policy with LLM-based Agent), which integrates Large Language Models (LLMs) into economic simulations, aiming to bridge the gap between theoretical models and real-world dynamics. Using public housing allocation scenarios as a case study, we conduct extensive policy simulation experiments to verify the feasibility and effectiveness of the SRAP-Agent and employ the Policy Optimization Algorithm with certain optimization objectives. The source code can be found in https://github.com/jijiarui-cather/SRAPAgent_Framework
LAMPER: LanguAge Model and Prompt EngineeRing for zero-shot time series classification
Du, Zhicheng, Xie, Zhaotian, Tong, Yan, Qin, Peiwu
This study constructs the LanguAge Model with Prompt EngineeRing (LAMPER) framework, designed to systematically evaluate the adaptability of pre-trained language models (PLMs) in accommodating diverse prompts and their integration in zero-shot time series (TS) classification. Our findings indicate that the feature representation capacity of LAMPER is influenced by the maximum input token threshold imposed by PLMs. The exploration of time series (TS)-based tasks constitutes a research-intensive domain with significant implications with wide-ranging implications in diverse professional fields, including healthcare, finance, and energy (Zhang et al., 2022; Zheng et al., 2023; Santoro et al., 2023). Within the realms of natural language processing (NLP), the dynamic landscape witnesses the rapid evolution of pre-trained language models (PLMs) and prompt engineering (Min et al., 2023; Wei et al., 2022). These advancements underscore their commendable capacity to adeptly execute an extensive array of tasks, particularly under few-shot or even zero-shot conditions (Brown et al., 2020; Webson & Pavlick, 2022).
Cognitive resilience: Unraveling the proficiency of image-captioning models to interpret masked visual content
Du, Zhicheng, Xie, Zhaotian, Ying, Huazhang, Zhang, Likun, Qin, Peiwu
This study explores the ability of Image Captioning (IC) models to decode masked visual content sourced from diverse datasets. Our findings reveal the IC model's capability to generate captions from masked images, closely resembling the original content. Notably, even in the presence of masks, the model adeptly crafts descriptive textual information that goes beyond what is observable in the original image-generated captions. While the decoding performance of the IC model experiences a decline with an increase in the masked region's area, the model still performs well when important regions of the image are not masked at high coverage.
GAME: Generalized deep learning model towards multimodal data integration for early screening of adolescent mental disorders
Du, Zhicheng, Jiang, Chenyao, Yuan, Xi, Zhai, Shiyao, Lei, Zhengyang, Ma, Shuyue, Liu, Yang, Ye, Qihui, Xiao, Chufan, Huang, Qiming, Xu, Ming, Yu, Dongmei, Qin, Peiwu
The timely identification of mental disorders in adolescents is a global public health challenge.Single factor is difficult to detect the abnormality due to its complex and subtle nature. Additionally, the generalized multimodal Computer-Aided Screening (CAS) systems with interactive robots for adolescent mental disorders are not available. Here, we design an android application with mini-games and chat recording deployed in a portable robot to screen 3,783 middle school students and construct the multimodal screening dataset, including facial images, physiological signs, voice recordings, and textual transcripts.We develop a model called GAME (Generalized Model with Attention and Multimodal EmbraceNet) with novel attention mechanism that integrates cross-modal features into the model. GAME evaluates adolescent mental conditions with high accuracy (73.34%-92.77%) and F1-Score (71.32%-91.06%).We find each modality contributes dynamically to the mental disorders screening and comorbidities among various mental disorders, indicating the feasibility of explainable model. This study provides a system capable of acquiring multimodal information and constructs a generalized multimodal integration algorithm with novel attention mechanisms for the early screening of adolescent mental disorders.