Large Language Model as Attributed Training Data Generator: A T ale of Diversity and Bias Yue Y u

Neural Information Processing Systems 

Large language models (LLMs) have been recently leveraged as training data generators for various natural language processing (NLP) tasks. While previous research has explored different approaches to training models using generated data, they generally rely on simple class-conditional prompts, which may limit the diversity of the generated data and inherit systematic biases of LLM. Thus, we investigate training data generation with diversely attributed prompts (e.g.,

Similar Docs  Excel Report  more

TitleSimilaritySource
None found