ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.07077
  4. Cited By
Multi-turn Evaluation of Anthropomorphic Behaviours in Large Language Models
v1v2 (latest)

Multi-turn Evaluation of Anthropomorphic Behaviours in Large Language Models

10 February 2025
Lujain Ibrahim
Canfer Akbulut
Rasmi Elasmar
Charvi Rastogi
Minsuk Kahng
Meredith Ringel Morris
Kevin R. McKee
Verena Rieser
Murray Shanahan
Laura Weidinger
ArXiv (abs)PDFHTMLGithub (5★)

Papers citing "Multi-turn Evaluation of Anthropomorphic Behaviours in Large Language Models"

9 / 9 papers shown
Title
Generative Value Conflicts Reveal LLM Priorities
Generative Value Conflicts Reveal LLM Priorities
Andy Liu
Kshitish Ghate
Mona Diab
Daniel Fried
Atoosa Kasirzadeh
Max Kleiman-Weiner
96
1
0
29 Sep 2025
HumanAgencyBench: Scalable Evaluation of Human Agency Support in AI Assistants
HumanAgencyBench: Scalable Evaluation of Human Agency Support in AI Assistants
Benjamin Sturgeon
Daniel Samuelson
Jacob Haimes
Jacy Reese Anthis
139
1
0
10 Sep 2025
Measuring and mitigating overreliance is necessary for building human-compatible AI
Measuring and mitigating overreliance is necessary for building human-compatible AI
Lujain Ibrahim
Katherine M. Collins
Sunnie S. Y. Kim
Anka Reuel
Max Lamparth
...
Siddharth Swaroop
Ilia Sucholutsky
A. Strait
Q. V. Liao
Umang Bhatt
100
3
0
08 Sep 2025
The Personality Illusion: Revealing Dissociation Between Self-Reports & Behavior in LLMs
The Personality Illusion: Revealing Dissociation Between Self-Reports & Behavior in LLMs
Pengrui Han
Rafal Kocielnik
Peiyang Song
Ramit Debnath
Dean Mobbs
Anima Anandkumar
R. Alvarez
225
4
0
03 Sep 2025
BioBlue: Notable runaway-optimiser-like LLM failure modes on biologically and economically aligned AI safety benchmarks for LLMs with simplified observation format
BioBlue: Notable runaway-optimiser-like LLM failure modes on biologically and economically aligned AI safety benchmarks for LLMs with simplified observation format
Roland Pihlakas
Sruthi Kuriakose
32
0
0
02 Sep 2025
Training language models to be warm and empathetic makes them less reliable and more sycophantic
Training language models to be warm and empathetic makes them less reliable and more sycophantic
Lujain Ibrahim
Franziska Sofia Hafner
Luc Rocher
139
7
0
29 Jul 2025
Deflating Deflationism: A Critical Perspective on Debunking Arguments Against LLM Mentality
Deflating Deflationism: A Critical Perspective on Debunking Arguments Against LLM Mentality
Alex Grzankowski
Geoff Keeling
Henry Shevlin
Winnie Street
98
0
0
16 Jun 2025
Not Minds, but Signs: Reframing LLMs through Semiotics
Not Minds, but Signs: Reframing LLMs through Semiotics
Davide Picca
145
2
0
20 May 2025
Artificial Intelligence in Deliberation: The AI Penalty and the Emergence of a New Deliberative DivideGovernment Information Quarterly (GIQ), 2025
Andreas Jungherr
Adrian Rauchfleisch
224
4
0
10 Mar 2025
1