-
Aligning Human and LLM Judgments: Insights from EvalAssist on
Task-Specific Evaluations and AI-assisted Assessment Strategy
Preferences
Preprint – Read Paper -
EvalAssist: LLM-as-a-Judge Simplified
AAAI 2025 – Read Paper -
Black-box Uncertainty Quantification Method for
LLM-as-a-Judge
NeurIPS 2024 – Read Paper -
Human-Centered Design Recommendations for
LLM-as-a-Judge
ACL 2024 Workshop HuCLLM – Read Paper -
EvaluLLM: LLM Assisted Evaluation of Generative Outputs
IUI ’24 Companion – Read Paper