Papers
arxiv:2505.13157

Role-Playing Evaluation for Large Language Models

Published on May 19
· Submitted by yelboudouri on Jun 2
Authors:
,
,

Abstract

A benchmark called Role-Playing Eval assesses Large Language Models in role-playing across emotional understanding, decision-making, moral alignment, and in-character consistency.

AI-generated summary

Large Language Models (LLMs) demonstrate a notable capacity for adopting personas and engaging in role-playing. However, evaluating this ability presents significant challenges, as human assessments are resource-intensive and automated evaluations can be biased. To address this, we introduce Role-Playing Eval (RPEval), a novel benchmark designed to assess LLM role-playing capabilities across four key dimensions: emotional understanding, decision-making, moral alignment, and in-character consistency. This article details the construction of RPEval and presents baseline evaluations. Our code and dataset are available at https://github.com/yelboudouri/RPEval

Community

Paper author Paper submitter

Hey everyone,

We've put together a benchmark that evaluates LLMs based on their roleplaying capabilities. We're now building a leaderboard that includes evaluations of both open-source and proprietary models. So far, we've evaluated 8 different models using the RPEval method introduced in this paper.

If there's a specific model you'd like us to include, or if you have suggestions to improve the evaluation, feel free to share them!

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2505.13157 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2505.13157 in a Space README.md to link it from this page.

Collections including this paper 1