Goto

Collaborating Authors

 nlu



* says " It will be of great help to the improvement of the generalization ability of the

Neural Information Processing Systems

We thank each reviewer for taking the time to thoughtfully comment on our work and we're glad that they recognize NLU tasks, such as teaching autonomous agents to perform tasks by demonstration. R2 wonders: don't these results just show R2 also points out some things that are unclear in the experiments section. It's true the models also perform well on the random split (A), which we left unsaid but will Finally, we thank R2 for pointing out 2 missing links in Figure 1, we will update them accordingly. C can shed light on this. GECA adds a lot of red squares to the training set.


Enhancing Text Generation in Joint NLG/NLU Learning Through Curriculum Learning, Semi-Supervised Training, and Advanced Optimization Techniques

Shaik, Rahimanuddin, Kishore, Katikela Sreeharsha

arXiv.org Artificial Intelligence

Text generation is the automated process of producing written or spoken language using computational methods. It involves generating coherent and contextually relevant text based on predefined rules or learned patterns. However, challenges in text generation arise from maintaining coherence, ensuring diversity and creativity, and avoiding biases or inappropriate content. This research paper developed a novel approach to improve text generation in the context of joint Natural Language Generation (NLG) and Natural Language Understanding (NLU) learning. The data is prepared by gathering and preprocessing annotated datasets, including cleaning, tokenization, stemming, and stop-word removal. Feature extraction techniques such as POS tagging, Bag of words, and Term Frequency-Inverse Document Frequency (TF-IDF) are applied. Transformer-based encoders and decoders, capturing long range dependencies and improving source-target sequence modelling. Pre-trained language models like Optimized BERT are incorporated, along with a Hybrid Redfox Artificial Hummingbird Algorithm (HRAHA). Reinforcement learning with policy gradient techniques, semi-supervised training, improved attention mechanisms, and differentiable approximations like straight-through Gumbel SoftMax estimator are employed to fine-tune the models and handle complex linguistic tasks effectively. The proposed model is implemented using Python.


An Approach for Auto Generation of Labeling Functions for Software Engineering Chatbots

Alor, Ebube, Abdellatif, Ahmad, Khatoonabadi, SayedHassan, Shihab, Emad

arXiv.org Artificial Intelligence

Software engineering (SE) chatbots are increasingly gaining attention for their role in enhancing development processes. At the core of chatbots are the Natural Language Understanding platforms (NLUs), which enable them to comprehend and respond to user queries. Before deploying NLUs, there is a need to train them with labeled data. However, acquiring such labeled data for SE chatbots is challenging due to the scarcity of high-quality datasets. This challenge arises because training SE chatbots requires specialized vocabulary and phrases not found in typical language datasets. Consequently, chatbot developers often resort to manually annotating user queries to gather the data necessary for training effective chatbots, a process that is both time-consuming and resource-intensive. Previous studies propose approaches to support chatbot practitioners in annotating users' posed queries. However, these approaches require human intervention to generate rules, called labeling functions (LFs), that identify and categorize user queries based on specific patterns in the data. To address this issue, we propose an approach to automatically generate LFs by extracting patterns from labeled user queries. We evaluate the effectiveness of our approach by applying it to the queries of four diverse SE datasets (namely AskGit, MSA, Ask Ubuntu, and Stack Overflow) and measure the performance improvement gained from training the NLU on the queries labeled by the generated LFs. We find that the generated LFs effectively label data with AUC scores of up to 85.3%, and NLU's performance improvement of up to 27.2% across the studied datasets. Furthermore, our results show that the number of LFs used to generate LFs affects the labeling performance. We believe that our approach can save time and resources in labeling users' queries, allowing practitioners to focus on core chatbot functionalities.


SQATIN: Supervised Instruction Tuning Meets Question Answering for Improved Dialogue NLU

Razumovskaia, Evgeniia, Glavaš, Goran, Korhonen, Anna, Vulić, Ivan

arXiv.org Artificial Intelligence

Task-oriented dialogue (ToD) systems help users execute well-defined tasks across a variety of domains (e.g., $\textit{flight booking}$ or $\textit{food ordering}$), with their Natural Language Understanding (NLU) components being dedicated to the analysis of user utterances, predicting users' intents ($\textit{Intent Detection}$, ID) and extracting values for informational slots ($\textit{Value Extraction}$, VE). In most domains, labelled NLU data is scarce, making sample-efficient learning -- enabled with effective transfer paradigms -- paramount. In this work, we introduce SQATIN, a new framework for dialog NLU based on (i) instruction tuning and (ii) question-answering-based formulation of ID and VE tasks. According to the evaluation on established NLU benchmarks, SQATIN sets the new state of the art in dialogue NLU, substantially surpassing the performance of current models based on standard fine-tuning objectives in both in-domain training and cross-domain transfer. SQATIN yields particularly large performance gains in cross-domain transfer, owing to the fact that our QA-based instruction tuning leverages similarities between natural language descriptions of classes (i.e., slots and intents) across domains.


HICL: Hashtag-Driven In-Context Learning for Social Media Natural Language Understanding

Tan, Hanzhuo, Xu, Chunpu, Li, Jing, Zhang, Yuqun, Fang, Zeyang, Chen, Zeyu, Lai, Baohua

arXiv.org Artificial Intelligence

Natural language understanding (NLU) is integral to various social media applications. However, existing NLU models rely heavily on context for semantic learning, resulting in compromised performance when faced with short and noisy social media content. To address this issue, we leverage in-context learning (ICL), wherein language models learn to make inferences by conditioning on a handful of demonstrations to enrich the context and propose a novel hashtag-driven in-context learning (HICL) framework. Concretely, we pre-train a model #Encoder, which employs #hashtags (user-annotated topic labels) to drive BERT-based pre-training through contrastive learning. Our objective here is to enable #Encoder to gain the ability to incorporate topic-related semantic information, which allows it to retrieve topic-related posts to enrich contexts and enhance social media NLU with noisy contexts. To further integrate the retrieved context with the source text, we employ a gradient-based method to identify trigger terms useful in fusing information from both sources. For empirical studies, we collected 45M tweets to set up an in-context NLU benchmark, and the experimental results on seven downstream tasks show that HICL substantially advances the previous state-of-the-art results. Furthermore, we conducted extensive analyzes and found that: (1) combining source input with a top-retrieved post from #Encoder is more effective than using semantically similar posts; (2) trigger words can largely benefit in merging context from the source and retrieved posts.


DRAGON: A Dialogue-Based Robot for Assistive Navigation with Visual Language Grounding

Liu, Shuijing, Hasan, Aamir, Hong, Kaiwen, Wang, Runxuan, Chang, Peixin, Mizrachi, Zachary, Lin, Justin, McPherson, D. Livingston, Rogers, Wendy A., Driggs-Campbell, Katherine

arXiv.org Artificial Intelligence

Persons with visual impairments (PwVI) have difficulties understanding and navigating spaces around them. Current wayfinding technologies either focus solely on navigation or provide limited communication about the environment. Motivated by recent advances in visual-language grounding and semantic navigation, we propose DRAGON, a guiding robot powered by a dialogue system and the ability to associate the environment with natural language. By understanding the commands from the user, DRAGON is able to guide the user to the desired landmarks on the map, describe the environment, and answer questions from visual observations. Through effective utilization of dialogue, the robot can ground the user's free-form descriptions to landmarks in the environment, and give the user semantic information through spoken language. We conduct a user study with blindfolded participants in an everyday indoor environment. Our results demonstrate that DRAGON is able to communicate with the user smoothly, provide a good guiding experience, and connect users with their surrounding environment in an intuitive manner.


Can We Use Probing to Better Understand Fine-tuning and Knowledge Distillation of the BERT NLU?

Hościłowicz, Jakub, Sowański, Marcin, Czubowski, Piotr, Janicki, Artur

arXiv.org Artificial Intelligence

In this article, we use probing to investigate phenomena that occur during fine-tuning and knowledge distillation of a BERT-based natural language understanding (NLU) model. Our ultimate purpose was to use probing to better understand practical production problems and consequently to build better NLU models. We designed experiments to see how fine-tuning changes the linguistic capabilities of BERT, what the optimal size of the fine-tuning dataset is, and what amount of information is contained in a distilled NLU based on a tiny Transformer. The results of the experiments show that the probing paradigm in its current form is not well suited to answer such questions. Structural, Edge and Conditional probes do not take into account how easy it is to decode probed information. Consequently, we conclude that quantification of information decodability is critical for many practical applications of the probing paradigm.


Using Natural Language Processing to Uncover Valuable Insights in Text-based Data - insideBIGDATA

#artificialintelligence

In this special guest feature, Ryan Welsh, Co-founder and CEO of Kyndi, discusses how organizations are leveraging the latest natural language processing techniques to enable sophisticated natural language understanding. Ryan started Kyndi in 2014 with a vision of creating a world where AI would empower humans to do their most meaningful work. Under his leadership, Kyndi has created the natural language enablement category, offering a powerful Natural Language Enablement Platform and natural language-enabled solutions. Ryan received his B.A. in Anthropology from The Catholic University of America, his M.S. in Applied Math/Economics from Rutgers University, and an M.B.A. from the University of Notre Dame. According to Deloitte, as much as 80% of all information is hidden in unstructured, text-based data living in various systems inside and outside of the companies.


Adaptive Natural Language Generation for Task-oriented Dialogue via Reinforcement Learning

Ohashi, Atsumoto, Higashinaka, Ryuichiro

arXiv.org Artificial Intelligence

When a natural language generation (NLG) component is implemented in a real-world task-oriented dialogue system, it is necessary to generate not only natural utterances as learned on training data but also utterances adapted to the dialogue environment (e.g., noise from environmental sounds) and the user (e.g., users with low levels of understanding ability). Inspired by recent advances in reinforcement learning (RL) for language generation tasks, we propose ANTOR, a method for Adaptive Natural language generation for Task-Oriented dialogue via Reinforcement learning. In ANTOR, a natural language understanding (NLU) module, which corresponds to the user's understanding of system utterances, is incorporated into the objective function of RL. If the NLG's intentions are correctly conveyed to the NLU, which understands a system's utterances, the NLG is given a positive reward. We conducted experiments on the MultiWOZ dataset, and we confirmed that ANTOR could generate adaptive utterances against speech recognition errors and the different vocabulary levels of users.