ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2402.10645
  4. Cited By
Can Separators Improve Chain-of-Thought Prompting?

Can Separators Improve Chain-of-Thought Prompting?

16 February 2024
Yoonjeong Park
Hyunjin Kim
Chanyeol Choi
Junseong Kim
Jy-yong Sohn
    LRM
    ReLM
ArXivPDFHTML

Papers citing "Can Separators Improve Chain-of-Thought Prompting?"

4 / 4 papers shown
Title
Self-Convinced Prompting: Few-Shot Question Answering with Repeated
  Introspection
Self-Convinced Prompting: Few-Shot Question Answering with Repeated Introspection
Haodi Zhang
Min Cai
Xinhe Zhang
Chen Zhang
Rui Mao
Kaishun Wu
KELM
LRM
ReLM
17
8
0
08 Oct 2023
What Makes Pre-trained Language Models Better Zero-shot Learners?
What Makes Pre-trained Language Models Better Zero-shot Learners?
Jinghui Lu
Dongsheng Zhu
Weidong Han
Rui Zhao
Brian Mac Namee
Fei Tan
34
21
0
30 Sep 2022
Self-Consistency Improves Chain of Thought Reasoning in Language Models
Self-Consistency Improves Chain of Thought Reasoning in Language Models
Xuezhi Wang
Jason W. Wei
Dale Schuurmans
Quoc Le
Ed H. Chi
Sharan Narang
Aakanksha Chowdhery
Denny Zhou
ReLM
BDL
LRM
AI4CE
297
3,163
0
21 Mar 2022
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
Jason W. Wei
Xuezhi Wang
Dale Schuurmans
Maarten Bosma
Brian Ichter
F. Xia
Ed H. Chi
Quoc Le
Denny Zhou
LM&Ro
LRM
AI4CE
ReLM
315
8,261
0
28 Jan 2022
1