Goto

Collaborating Authors

 Zhang, Xinyun


Reconstruct Before Summarize: An Efficient Two-Step Framework for Condensing and Summarizing Meeting Transcripts

arXiv.org Artificial Intelligence

Based on this understanding, Although numerous achievements have been made we propose a two-step meeting summarization in the well-structured text abstractive summarization framework, Reconstrcut before Summarize(RbS), (Zhang et al., 2020a; Liu* et al., 2018; Lewis to address the challenge of scattered et al., 2020), the research on meeting summarization information in meetings. RbS adopts a reconstructor is still stretched in limit. There are some outstanding to reconstruct the responses in the meeting, it challenges in this field, including 1) much also synchronically traces out which texts in the noise brought from automated speech recognition meeting drove the responses and marks them as models; 2) lengthy meeting transcripts consisting essential contents. Therefore, salient information of casual conversations, content redundancy, and is captured and annotated as anchor tokens in RbS.


Towards Versatile and Efficient Visual Knowledge Integration into Pre-trained Language Models with Cross-Modal Adapters

arXiv.org Artificial Intelligence

Humans learn language via multi-modal knowledge. However, due to the text-only pre-training scheme, most existing pre-trained language models (PLMs) are hindered from the multi-modal information. To inject visual knowledge into PLMs, existing methods incorporate either the text or image encoder of vision-language models (VLMs) to encode the visual information and update all the original parameters of PLMs for knowledge fusion. In this paper, we propose a new plug-and-play module, X-adapter, to flexibly leverage the aligned visual and textual knowledge learned in pre-trained VLMs and efficiently inject them into PLMs. Specifically, we insert X-adapters into PLMs, and only the added parameters are updated during adaptation. To fully exploit the potential in VLMs, X-adapters consist of two sub-modules, V-expert and T-expert, to fuse VLMs' image and text representations, respectively. We can opt for activating different sub-modules depending on the downstream tasks. Experimental results show that our method can significantly improve the performance on object-color reasoning and natural language understanding (NLU) tasks compared with PLM baselines.


ChatEDA: A Large Language Model Powered Autonomous Agent for EDA

arXiv.org Artificial Intelligence

The integration of a complex set of Electronic Design Automation (EDA) tools to enhance interoperability is a critical concern for circuit designers. Recent advancements in large language models (LLMs) have showcased their exceptional capabilities in natural language processing and comprehension, offering a novel approach to interfacing with EDA tools. This research paper introduces ChatEDA, an autonomous agent for EDA empowered by a large language model, AutoMage, complemented by EDA tools serving as executors. ChatEDA streamlines the design flow from the Register-Transfer Level (RTL) to the Graphic Data System Version II (GDSII) by effectively managing task planning, script generation, and task execution. Through comprehensive experimental evaluations, ChatEDA has demonstrated its proficiency in handling diverse requirements, and our fine-tuned AutoMage model has exhibited superior performance compared to GPT-4 and other similar LLMs.