Goto

Collaborating Authors

 nlg


Natural Language Generation in Healthcare: A Review of Methods and Applications

Lyu, Mengxian, Li, Xiaohan, Chen, Ziyi, Pan, Jinqian, Peng, Cheng, Talankar, Sankalp, Wu, Yonghui

arXiv.org Artificial Intelligence

Natural language generation (NLG) is the key technology to achieve generative artificial intelligence (AI). With the breakthroughs in large language models (LLMs), NLG has been widely used in various medical applications, demonstrating the potential to enhance clinical workflows, support clinical decision-making, and improve clinical documentation. Heterogeneous and diverse medical data modalities, such as medical text, images, and knowledge bases, are utilized in NLG. Researchers have proposed many generative models and applied them in a number of healthcare applications. There is a need for a comprehensive review of NLG methods and applications in the medical domain. In this study, we systematically reviewed 113 scientific publications from a total of 3,988 NLG-related articles identified using a literature search, focusing on data modality, model architecture, clinical applications, and evaluation methods. Following PRISMA (Preferred Reporting Items for Systematic reviews and Meta-Analyses) guidelines, we categorize key methods, identify clinical applications, and assess their capabilities, limitations, and emerging challenges. This timely review covers the key NLG technologies and medical applications and provides valuable insights for future studies to leverage NLG to transform medical discovery and healthcare.


Beyond Generative Artificial Intelligence: Roadmap for Natural Language Generation

Maestre, María Miró, Martínez-Murillo, Iván, Martin, Tania J., Navarro-Colorado, Borja, Ferrández, Antonio, Cueto, Armando Suárez, Lloret, Elena

arXiv.org Artificial Intelligence

Generative Artificial Intelligence has grown exponentially as a result of Large Language Models (LLMs). This has been possible because of the impressive performance of deep learning methods created within the field of Natural Language Processing (NLP) and its subfield Natural Language Generation (NLG), which is the focus of this paper. Within the growing LLM family are the popular GPT-4, Bard and more specifically, tools such as ChatGPT have become a benchmark for other LLMs when solving most of the tasks involved in NLG research. This scenario poses new questions about the next steps for NLG and how the field can adapt and evolve to deal with new challenges in the era of LLMs. To address this, the present paper conducts a review of a representative sample of surveys recently published in NLG. By doing so, we aim to provide the scientific community with a research roadmap to identify which NLG aspects are still not suitably addressed by LLMs, as well as suggest future lines of research that should be addressed going forward.


Many Hands Make Light Work: Task-Oriented Dialogue System with Module-Based Mixture-of-Experts

Su, Ruolin, Juang, Biing-Hwang

arXiv.org Artificial Intelligence

Task-oriented dialogue systems are broadly used in virtual assistants and other automated services, providing interfaces between users and machines to facilitate specific tasks. Nowadays, task-oriented dialogue systems have greatly benefited from pre-trained language models (PLMs). However, their task-solving performance is constrained by the inherent capacities of PLMs, and scaling these models is expensive and complex as the model size becomes larger. To address these challenges, we propose Soft Mixture-of-Expert Task-Oriented Dialogue system (SMETOD) which leverages an ensemble of Mixture-of-Experts (MoEs) to excel at subproblems and generate specialized outputs for task-oriented dialogues. SMETOD also scales up a task-oriented dialogue system with simplicity and flexibility while maintaining inference efficiency. We extensively evaluate our model on three benchmark functionalities: intent prediction, dialogue state tracking, and dialogue response generation. Experimental results demonstrate that SMETOD achieves state-of-the-art performance on most evaluated metrics. Moreover, comparisons against existing strong baselines show that SMETOD has a great advantage in the cost of inference and correctness in problem-solving.


Self-training from Self-memory in Data-to-text Generation

Ta, Hoang-Thang

arXiv.org Artificial Intelligence

This paper introduces a novel training model, self-training from self-memory (STSM) in data-to-text generation (DTG), allowing the model to self-train on subsets, including self-memory as outputs inferred directly from the trained models and/or the new data. The quality of self-memory is validated by two models, data-to-text (D2T) and text-to-data (T2D), by two pre-defined conditions: (1) the appearance of all source values in the outputs of the D2T model and (2) the ability to convert back to source data in the outputs in the T2D model. We utilize a greedy algorithm to generate shorter D2T outputs if they contain all source values. Subsequently, we use the T2D model to confirm that these outputs can capture input relationships by demonstrating their capacity to convert text back into data. With 30% of the dataset, we can train the D2T model with a competitive performance compared to full training in the same setup. We experiment with our model on two datasets, E2E NLG and DART. STSM offers the D2T model a generalization capability from its subset memory while reducing training data volume. Ultimately, we anticipate that this paper will contribute to continual learning solutions that adapt to new training data, incorporating it as a form of self-memory in DTG tasks. The curated dataset is publicly available at: https://github.com/hoangthangta/STSM.


Information Filter upon Diversity-Improved Decoding for Diversity-Faithfulness Tradeoff in NLG

Meng, Han, He, Xiaosong, Chen, Zexing, Zhou, Feng

arXiv.org Artificial Intelligence

Some Natural Language Generation (NLG) tasks require both faithfulness and diversity. The decoding strategy is intensively related to the quality of the generated text. Strategies such as beam search, greedy search, etc., perform with low diversity and high repetition. On the other hand, guided decoding, the solution towards diversity, may generate unfaithful expressions. To this end, this paper presents Information Filter upon Diversity-Improved Decoding (IFDID) to obtain the tradeoff between diversity and faithfulness. IFDID is a two-stage decoding strategy leveraging the proposed Enhance-Filter framework, which achieves the tradeoff by increasing the probabilities of some typical tokens being selected and subsequently filtering them by their information amount. To verify the effectiveness, we compare our method with other baselines on related CommonGEN, RocStories and AdGen benchmarks, which cover Chinese and English datasets. Our numerical experimental results and human evaluation outcomes verify the effectiveness of the proposed approach, as our approach achieves a 1.24 higher ROUGE score describing faithfulness as well as higher diversity represented by 62.5% higher upon Dist-2 than traditional approaches, demonstrating that IFDID is a novel SOTA decoding strategy for the tradeoff between diversity and faithfulness.


Adaptive Natural Language Generation for Task-oriented Dialogue via Reinforcement Learning

Ohashi, Atsumoto, Higashinaka, Ryuichiro

arXiv.org Artificial Intelligence

When a natural language generation (NLG) component is implemented in a real-world task-oriented dialogue system, it is necessary to generate not only natural utterances as learned on training data but also utterances adapted to the dialogue environment (e.g., noise from environmental sounds) and the user (e.g., users with low levels of understanding ability). Inspired by recent advances in reinforcement learning (RL) for language generation tasks, we propose ANTOR, a method for Adaptive Natural language generation for Task-Oriented dialogue via Reinforcement learning. In ANTOR, a natural language understanding (NLU) module, which corresponds to the user's understanding of system utterances, is incorporated into the objective function of RL. If the NLG's intentions are correctly conveyed to the NLU, which understands a system's utterances, the NLG is given a positive reward. We conducted experiments on the MultiWOZ dataset, and we confirmed that ANTOR could generate adaptive utterances against speech recognition errors and the different vocabulary levels of users.


NLP, NLU, and NLG

#artificialintelligence

Images used in my articles are Properties of the Respective Organisations and are used here solely for Reference, Illustrative and Educational Purposes Only. We've understood what Natural Language Processing, (NLP) is right? But we haven't understood much about what Natural Language Understanding (NLU), and Natural Language Generation (NLG) are. Let's understand both of them, along with NLP in this article. Natural Language Processing (NLP) refers to the branch of Computer Science, and more specifically, the branch of Artificial Intelligence (AI)concerned with giving computers the ability to understand the text and spoken words in much the same way human beings can.


NLP, NLU, and NLG: What's The Difference? A Comprehensive Guide - KDnuggets

#artificialintelligence

Have you ever used a smart assistant (think something like Siri or Alexa) to answer questions for you? The answer is more than likely "yes", which means that you are, on some level, already familiar with what's known as natural language processing (NLP). NLP is the combination of methods taken from different disciplines that smart assistants like Siri and Alexa use to make sense of the questions we ask them. It combines disciplines such as artificial intelligence and computer science to make it easier for human beings to talk with computers the way we would with another person. This idea of having a facsimile of a human conversation with a machine goes back to a groundbreaking paper written by Alan Turing -- a paper that formed the basis for NLP technology that we use today.


Natural Language Generation (NLG): Everything You Need to Know

#artificialintelligence

Chatbots, voice assistants, and AI blog writers (to name a few) all use natural language generation. NLG systems can turn numbers into narratives based on pre-set templates. They can predict which words need to be generated next (in, say, an email you're actively typing). Or, the most sophisticated systems can formulate entire summaries, articles, or responses.


NLP for Banking & Financial Services

#artificialintelligence

Every large financial services company is pursuing AI initiatives in 2022. Some are in the earliest stages. Others have been investing in the technology for years. Regardless, all recognize that natural-language processing (NLP) is a foundational capability for their long-term AI business goals. Unfortunately NLP is still an emerging technology, and the best options for deploying it are not immediately obvious.