🏢 City University of Hong Kong
Crowd Comparative Reasoning: Unlocking Comprehensive Evaluations for LLM-as-a-Judge
·3819 words·18 mins·
loading
·
loading
AI Generated
🤗 Daily Papers
Natural Language Processing
Large Language Models
🏢 City University of Hong Kong
Crowd-based comparative evaluation significantly boosts LLM-as-a-judge accuracy by using crowd responses to expose deeper details, resulting in more reliable and efficient auto-evaluation.