We gratefully acknowledge support from
the Simons Foundation and member institutions.
Full-text links:

Download:

Current browse context:

cs.CL

Change to browse by:

References & Citations

DBLP - CS Bibliography

Bookmark

(what is this?)
CiteULike logo BibSonomy logo Mendeley logo del.icio.us logo Digg logo Reddit logo

Computer Science > Computation and Language

Title: Towards better Human-Agent Alignment: Assessing Task Utility in LLM-Powered Applications

Abstract: The rapid development in the field of Large Language Models (LLMs) has led to a surge in applications that facilitate collaboration among multiple agents to assist humans in their daily tasks. However, a significant gap remains in assessing whether LLM-powered applications genuinely enhance user experience and task execution efficiency. This highlights the pressing need for methods to verify utility of LLM-powered applications, particularly by ensuring alignment between the application's functionality and end-user needs. We introduce AgentEval provides an implementation for the math problems, a novel framework designed to simplify the utility verification process by automatically proposing a set of criteria tailored to the unique purpose of any given application. This allows for a comprehensive assessment, quantifying the utility of an application against the suggested criteria. We present a comprehensive analysis of the robustness of quantifier's work.
Subjects: Computation and Language (cs.CL); Artificial Intelligence (cs.AI)
Cite as: arXiv:2402.09015 [cs.CL]
  (or arXiv:2402.09015v3 [cs.CL] for this version)

Submission history

From: Julia Kiseleva [view email]
[v1] Wed, 14 Feb 2024 08:46:15 GMT (14525kb,D)
[v2] Thu, 15 Feb 2024 18:24:03 GMT (15116kb,D)
[v3] Thu, 22 Feb 2024 23:49:10 GMT (15117kb,D)

Link back to: arXiv, form interface, contact.