Goto

Collaborating Authors

 cnt




Appendix to TreeMoCo: Contrastive Neuron Morphology Representation Learning

Neural Information Processing Systems

The formulas of these two metrics are detailed in Sec. Figure A1 (a), by only re-ordering of edge vectors, a branch's shape can be significantly changed Such measurement is taken in a bag-of-word (BOW) fashion. The context and spatial relationship between arbors are ignored under this measurement. Branch A and B are composed of the same set of edge vectors. Each neuron is viewed from three different angles. Three public datasets are used in this study.


Exploiting Symbolic Heuristics for the Synthesis of Domain-Specific Temporal Planning Guidance using Reinforcement Learning

Brugnara, Irene, Valentini, Alessandro, Micheli, Andrea

arXiv.org Artificial Intelligence

Recent work investigated the use of Reinforcement Learning (RL) for the synthesis of heuristic guidance to improve the performance of temporal planners when a domain is fixed and a set of training problems (not plans) is given. The idea is to extract a heuristic from the value function of a particular (possibly infinite-state) MDP constructed over the training problems. In this paper, we propose an evolution of this learning and planning framework that focuses on exploiting the information provided by symbolic heuristics during both the RL and planning phases. First, we formalize different reward schemata for the synthesis and use symbolic heuristics to mitigate the problems caused by the truncation of episodes needed to deal with the potentially infinite MDP . Second, we propose learning a residual of an existing symbolic heuristic, which is a "correction" of the heuristic value, instead of eagerly learning the whole heuristic from scratch. Finally, we use the learned heuristic in combination with a symbolic heuristic using a multiple-queue planning approach to balance systematic search with imperfect learned information. We experimentally compare all the approaches, highlighting their strengths and weaknesses and significantly advancing the state of the art for this planning and learning schema.


Universal Narrative Model: an Author-centric Storytelling Framework for Generative AI

Gerba, Hank

arXiv.org Artificial Intelligence

In their survey of authoring tools for computational narrative, Kybartas and Bidarra note that "we believe that creating a standard model of computational narrative could allow different systems to interact with the same narrative, without being restricted by incompatible models and definitions. Furthermore, such a model would also facilitate research into the generation of specific story components, e.g., allowing for multiple generators and even authors to collaborate on a given narrative" [Kybartas and Bidarra [2017]]. This paper proposes such a standard: the Universal Narrative Model (UNM). We foresee that generative AI will enable a new paradigm of storytelling technologies and processes: from assisting a writer of linear media (novels, film, television, etc.) by allowing them to test out scenes and characters before committing them to a script, all the way through to real-time storytelling systems in videogames which respond to a player's agency, and countless use cases in between [Peng et al. [2024]]. The UNM is designed to service any use case in which coherent narrative structure is a consideration, and in which authorial intent and direction is privileged. In the last five years, a robust body of research has demonstrated a wide variety of potential uses for computational narrative systems powered by generative AI, and some limited commercial deployments already exist [Yang et al. [2024], Hu et al. [2024]]. With such promise, however, comes a series of challenges: technical, narrative, and ethical. The goal of the Entertainment Technology Center's "Universal Narrative Model" project was to produce the UNM as an open standard. The ultimate directive of the project was to privilege, above all else, author-centric design and functionality, setting the stage for generative workflows which extend an author's narrative intent and creativity, rather than eclipse or replace it.


Enhancing Large Language Models for Hardware Verification: A Novel SystemVerilog Assertion Dataset

Menon, Anand, Miftah, Samit S, Kundu, Shamik, Kundu, Souvik, Srivastava, Amisha, Raha, Arnab, Sonnenschein, Gabriel Theodor, Banerjee, Suvadeep, Mathaikutty, Deepak, Basu, Kanad

arXiv.org Artificial Intelligence

Hardware verification is crucial in modern SoC design, consuming around 70% of development time. SystemVerilog assertions ensure correct functionality. However, existing industrial practices rely on manual efforts for assertion generation, which becomes increasingly untenable as hardware systems become complex. Recent research shows that Large Language Models (LLMs) can automate this process. However, proprietary SOTA models like GPT-4o often generate inaccurate assertions and require expensive licenses, while smaller open-source LLMs need fine-tuning to manage HDL code complexities. To address these issues, we introduce **VERT**, an open-source dataset designed to enhance SystemVerilog assertion generation using LLMs. VERT enables researchers in academia and industry to fine-tune open-source models, outperforming larger proprietary ones in both accuracy and efficiency while ensuring data privacy through local fine-tuning and eliminating costly licenses. The dataset is curated by systematically augmenting variables from open-source HDL repositories to generate synthetic code snippets paired with corresponding assertions. Experimental results demonstrate that fine-tuned models like Deepseek Coder 6.7B and Llama 3.1 8B outperform GPT-4o, achieving up to 96.88% improvement over base models and 24.14% over GPT-4o on platforms including OpenTitan, CVA6, OpenPiton and Pulpissimo. VERT is available at https://github.com/AnandMenon12/VERT.


Dissertation Machine Learning in Materials Science -- A case study in Carbon Nanotube field effect transistors

Tan, Shulin

arXiv.org Artificial Intelligence

Carbon Nanotube has long been seen as a promising candidate for high-performance electronic material, yet its unique 1D structure leads to challenges in device fabrication. Many processing approaches have been proposed to produce better performing CNTFETs and this explosion of data needs an efficient way to explore.


DR-RAG: Applying Dynamic Document Relevance to Retrieval-Augmented Generation for Question-Answering

Hei, Zijian, Liu, Weiling, Ou, Wenjie, Qiao, Juyi, Jiao, Junming, Song, Guowen, Tian, Ting, Lin, Yi

arXiv.org Artificial Intelligence

Retrieval-Augmented Generation (RAG) has recently demonstrated the performance of Large Language Models (LLMs) in the knowledge-intensive tasks such as Question-Answering (QA). RAG expands the query context by incorporating external knowledge bases to enhance the response accuracy. However, it would be inefficient to access LLMs multiple times for each query and unreliable to retrieve all the relevant documents by a single query. We have found that even though there is low relevance between some critical documents and query, it is possible to retrieve the remaining documents by combining parts of the documents with the query. To mine the relevance, a two-stage retrieval framework called Dynamic-Relevant Retrieval-Augmented Generation (DR-RAG) is proposed to improve document retrieval recall and the accuracy of answers while maintaining efficiency. Additionally, a compact classifier is applied to two different selection strategies to determine the contribution of the retrieved documents to answering the query and retrieve the relatively relevant documents. Meanwhile, DR-RAG call the LLMs only once, which significantly improves the efficiency of the experiment. The experimental results on multi-hop QA datasets show that DR-RAG can significantly improve the accuracy of the answers and achieve new progress in QA systems.


Statistical Learning under Heterogeneous Distribution Shift

Simchowitz, Max, Ajay, Anurag, Agrawal, Pulkit, Krishnamurthy, Akshay

arXiv.org Machine Learning

This paper studies the prediction of a target $\mathbf{z}$ from a pair of random variables $(\mathbf{x},\mathbf{y})$, where the ground-truth predictor is additive $\mathbb{E}[\mathbf{z} \mid \mathbf{x},\mathbf{y}] = f_\star(\mathbf{x}) +g_{\star}(\mathbf{y})$. We study the performance of empirical risk minimization (ERM) over functions $f+g$, $f \in F$ and $g \in G$, fit on a given training distribution, but evaluated on a test distribution which exhibits covariate shift. We show that, when the class $F$ is "simpler" than $G$ (measured, e.g., in terms of its metric entropy), our predictor is more resilient to heterogeneous covariate shifts} in which the shift in $\mathbf{x}$ is much greater than that in $\mathbf{y}$. Our analysis proceeds by demonstrating that ERM behaves qualitatively similarly to orthogonal machine learning: the rate at which ERM recovers the $f$-component of the predictor has only a lower-order dependence on the complexity of the class $G$, adjusted for partial non-indentifiability introduced by the additive structure. These results rely on a novel H\"older style inequality for the Dudley integral which may be of independent interest. Moreover, we corroborate our theoretical findings with experiments demonstrating improved resilience to shifts in "simpler" features across numerous domains.


MAILEX: Email Event and Argument Extraction

Srivastava, Saurabh, Singh, Gaurav, Matsumoto, Shou, Raz, Ali, Costa, Paulo, Poore, Joshua, Yao, Ziyu

arXiv.org Artificial Intelligence

In this work, we present the first dataset, MailEx, for performing event extraction from conversational email threads. To this end, we first proposed a new taxonomy covering 10 event types and 76 arguments in the email domain. Our final dataset includes 1.5K email threads and ~4K emails, which are annotated with totally ~8K event instances. To understand the task challenges, we conducted a series of experiments comparing three types of approaches, i.e., fine-tuned sequence labeling, fine-tuned generative extraction, and few-shot in-context learning. Our results showed that the task of email event extraction is far from being addressed, due to challenges lying in, e.g., extracting non-continuous, shared trigger spans, extracting non-named entity arguments, and modeling the email conversational history. Our work thus suggests more future investigations in this domain-specific event extraction task.