LLM Evaluation
MILE-RefHumEval: Multi-LLM Framework for Human-Aligned AI Evaluat
New research introduces a reference-free evaluation framework using multiple independent LLMs to assess AI outputs with better human alignment than single-judge approaches.