Moving LLM evaluation forward: lessons from human judgment research

This paper outlines a path toward more reliable and effective evaluation of Large Language Models (LLMs). It argues that insights from the study of human judgment and decision-making can illuminate current challenges in LLM assessment and help close critical gaps in how models are evaluated. By draw...

Full description

Saved in:
Bibliographic Details
Main Author: Andrea Polonioli
Format: Article
Language:English
Published: Frontiers Media S.A. 2025-05-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/frai.2025.1592399/full
Tags: Add Tag
No Tags, Be the first to tag this record!