Factuality-Aware Alignment for Large Language Models

Neural Information Processing Systems 

This makes SFT less factual as it trains on human-labeled data that may be novel to the LLM. Furthermore, reward functions used in standard RL often inadequately capture factuality and favor longer and more detailed responses, which inadvertently promote hallucination.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found