Rating Roulette: Self-Inconsistency in LLM-As-A-Judge Frameworks
Haldar, Rajarshi, Hockenmaier, Julia
–arXiv.org Artificial Intelligence
As Natural Language Generation (NLG) continues to be widely adopted, properly assessing it has become quite difficult. Lately, using large language models (LLMs) for evaluating these generations has gained traction, as they tend to align more closely with human preferences than conventional n-gram or embedding-based metrics. In our experiments, we show that LLM judges have low intra-rater reliability in their assigned scores across different runs. This variance makes their ratings inconsistent, almost arbitrary in the worst case, making it difficult to measure how good their judgments actually are. We quantify this inconsistency across different NLG tasks and benchmarks and see if judicious use of LLM judges can still be useful following proper guidelines.
arXiv.org Artificial Intelligence
Nov-3-2025
- Country:
- Asia
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- Singapore (0.04)
- Myanmar > Tanintharyi Region
- Europe
- North America
- Canada > Ontario
- Toronto (0.04)
- United States
- Illinois > Champaign County
- Urbana (0.04)
- Michigan > Washtenaw County
- Ann Arbor (0.04)
- Pennsylvania (0.04)
- Illinois > Champaign County
- Canada > Ontario
- Oceania > Australia
- New South Wales (0.04)
- Victoria > Melbourne (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.88)
- Industry:
- Health & Medicine (0.67)
- Leisure & Entertainment (0.46)
- Technology: