ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2308.09138
  4. Cited By
Semantic Consistency for Assuring Reliability of Large Language Models

Semantic Consistency for Assuring Reliability of Large Language Models

17 August 2023
Harsh Raj
Vipul Gupta
Domenic Rosati
S. Majumdar
    HILM
ArXivPDFHTML

Papers citing "Semantic Consistency for Assuring Reliability of Large Language Models"

13 / 13 papers shown
Title
Consistency in Language Models: Current Landscape, Challenges, and Future Directions
Consistency in Language Models: Current Landscape, Challenges, and Future Directions
Jekaterina Novikova
Carol Anderson
Borhane Blili-Hamelin
Subhabrata Majumdar
HILM
40
0
0
01 May 2025
Selective Prompt Anchoring for Code Generation
Selective Prompt Anchoring for Code Generation
Yuan Tian
Tianyi Zhang
56
3
0
24 Feb 2025
Benchmarking Prompt Sensitivity in Large Language Models
Benchmarking Prompt Sensitivity in Large Language Models
Amirhossein Razavi
Mina Soltangheis
Negar Arabzadeh
Sara Salamat
Morteza Zihayat
Ebrahim Bagheri
38
1
0
09 Feb 2025
A Statistical Analysis of LLMs' Self-Evaluation Using Proverbs
A Statistical Analysis of LLMs' Self-Evaluation Using Proverbs
Ryosuke Sonoda
Ramya Srinivasan
29
1
0
22 Oct 2024
Logically Consistent Language Models via Neuro-Symbolic Integration
Logically Consistent Language Models via Neuro-Symbolic Integration
Diego Calanzone
Stefano Teso
Antonio Vergari
LRM
32
6
0
09 Sep 2024
BeHonest: Benchmarking Honesty in Large Language Models
BeHonest: Benchmarking Honesty in Large Language Models
Steffi Chern
Zhulin Hu
Yuqing Yang
Ethan Chern
Yuan Guo
Jiahe Jin
Binjie Wang
Pengfei Liu
HILM
ALM
45
1
0
19 Jun 2024
INSIDE: LLMs' Internal States Retain the Power of Hallucination
  Detection
INSIDE: LLMs' Internal States Retain the Power of Hallucination Detection
Chao Chen
Kai-Chun Liu
Ze Chen
Yi Gu
Yue-bo Wu
Mingyuan Tao
Zhihang Fu
Jieping Ye
HILM
35
79
0
06 Feb 2024
Assessing the Reliability of Large Language Model Knowledge
Assessing the Reliability of Large Language Model Knowledge
Weixuan Wang
Barry Haddow
Alexandra Birch
Wei Peng
KELM
HILM
28
14
0
15 Oct 2023
Self-Consistency Improves Chain of Thought Reasoning in Language Models
Self-Consistency Improves Chain of Thought Reasoning in Language Models
Xuezhi Wang
Jason W. Wei
Dale Schuurmans
Quoc Le
Ed H. Chi
Sharan Narang
Aakanksha Chowdhery
Denny Zhou
ReLM
BDL
LRM
AI4CE
249
2,029
0
21 Mar 2022
BeliefBank: Adding Memory to a Pre-Trained Language Model for a
  Systematic Notion of Belief
BeliefBank: Adding Memory to a Pre-Trained Language Model for a Systematic Notion of Belief
Nora Kassner
Oyvind Tafjord
Hinrich Schütze
Peter Clark
KELM
LRM
199
51
0
29 Sep 2021
Measuring and Improving Consistency in Pretrained Language Models
Measuring and Improving Consistency in Pretrained Language Models
Yanai Elazar
Nora Kassner
Shauli Ravfogel
Abhilasha Ravichander
Eduard H. Hovy
Hinrich Schütze
Yoav Goldberg
HILM
234
273
0
01 Feb 2021
Language Models as Knowledge Bases?
Language Models as Knowledge Bases?
Fabio Petroni
Tim Rocktaschel
Patrick Lewis
A. Bakhtin
Yuxiang Wu
Alexander H. Miller
Sebastian Riedel
KELM
AI4MH
371
2,216
0
03 Sep 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
  Understanding
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
270
6,003
0
20 Apr 2018
1