ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2403.19305
  4. Cited By
MATEval: A Multi-Agent Discussion Framework for Advancing Open-Ended
  Text Evaluation
v1v2 (latest)

MATEval: A Multi-Agent Discussion Framework for Advancing Open-Ended Text Evaluation

28 March 2024
Yu Li
Shenyu Zhang
Rui Wu
Xiutian Huang
Yongrui Chen
Wenhao Xu
Guilin Qi
Dehai Min
    LLMAG
ArXiv (abs)PDFHTML

Papers citing "MATEval: A Multi-Agent Discussion Framework for Advancing Open-Ended Text Evaluation"

10 / 10 papers shown
Multi-Agent-as-Judge: Aligning LLM-Agent-Based Automated Evaluation with Multi-Dimensional Human Evaluation
Multi-Agent-as-Judge: Aligning LLM-Agent-Based Automated Evaluation with Multi-Dimensional Human Evaluation
Jiaju Chen
Yuxuan Lu
Xiaojie Wang
Huimin Zeng
Jing Huang
Jiri Gesi
Ying Xu
Bingsheng Yao
Dakuo Wang
LLMAGELM
215
9
0
28 Jul 2025
HiMATE: A Hierarchical Multi-Agent Framework for Machine Translation Evaluation
HiMATE: A Hierarchical Multi-Agent Framework for Machine Translation Evaluation
Shijie Zhang
Renhao Li
Songsheng Wang
Philipp Koehn
Min Yang
Derek F. Wong
181
2
0
22 May 2025
SeriesBench: A Benchmark for Narrative-Driven Drama Series Understanding
SeriesBench: A Benchmark for Narrative-Driven Drama Series UnderstandingComputer Vision and Pattern Recognition (CVPR), 2025
Yiming Lei
Chenkai Zhang
Ziqiang Liu
Haitao Leng
Shaoguo Liu
Tingting Gao
Qingjie Liu
Yunhong Wang
AI4TS
521
0
0
30 Apr 2025
Toward Generalizable Evaluation in the LLM Era: A Survey Beyond Benchmarks
Toward Generalizable Evaluation in the LLM Era: A Survey Beyond Benchmarks
Yixin Cao
Shibo Hong
Xuzhao Li
Jiahao Ying
Yubo Ma
...
Juanzi Li
Aixin Sun
Qi Zhang
Tat-Seng Chua
Tianwei Zhang
ALMELM
548
23
0
26 Apr 2025
Leveraging LLMs as Meta-Judges: A Multi-Agent Framework for Evaluating LLM Judgments
Leveraging LLMs as Meta-Judges: A Multi-Agent Framework for Evaluating LLM Judgments
Jian Wang
Jama Hussein Mohamud
Chongren Sun
Di Wu
Benoit Boulet
LLMAGELM
394
6
0
23 Apr 2025
Reading between the Lines: Can LLMs Identify Cross-Cultural Communication Gaps?
Reading between the Lines: Can LLMs Identify Cross-Cultural Communication Gaps?
Sougata Saha
Saurabh Kumar Pandey
Harshit Gupta
Monojit Choudhury
313
1
0
21 Feb 2025
Is my Meeting Summary Good? Estimating Quality with a Multi-LLM
  Evaluator
Is my Meeting Summary Good? Estimating Quality with a Multi-LLM EvaluatorInternational Conference on Computational Linguistics (COLING), 2024
Frederic Kirstein
Terry Ruas
Bela Gipp
356
3
0
27 Nov 2024
From Generation to Judgment: Opportunities and Challenges of LLM-as-a-judge
From Generation to Judgment: Opportunities and Challenges of LLM-as-a-judge
Dawei Li
Bohan Jiang
Liangjie Huang
Alimohammad Beigi
Chengshuai Zhao
...
Canyu Chen
Tianhao Wu
Kai Shu
Lu Cheng
Huan Liu
ELMAILaw
1.1K
287
0
25 Nov 2024
SAGEval: The frontiers of Satisfactory Agent based NLG Evaluation for
  reference-free open-ended text
SAGEval: The frontiers of Satisfactory Agent based NLG Evaluation for reference-free open-ended text
Reshmi Ghosh
Tianyi Yao
Lizzy Chen
Sadid Hasan
Tianwei Chen
Dario Bernal
Huitian Jiao
H M Sajjad Hossain
ELM
230
1
0
25 Nov 2024
What Makes a Good Story and How Can We Measure It? A Comprehensive
  Survey of Story Evaluation
What Makes a Good Story and How Can We Measure It? A Comprehensive Survey of Story Evaluation
Dingyi Yang
Qin Jin
407
15
0
26 Aug 2024
1