ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.14197
  4. Cited By
On the Representational Capacity of Neural Language Models with Chain-of-Thought Reasoning

On the Representational Capacity of Neural Language Models with Chain-of-Thought Reasoning

20 June 2024
Franz Nowak
Anej Svete
Alexandra Butoi
Ryan Cotterell
    ReLM
    LRM
ArXivPDFHTML

Papers citing "On the Representational Capacity of Neural Language Models with Chain-of-Thought Reasoning"

12 / 12 papers shown
Title
Unique Hard Attention: A Tale of Two Sides
Unique Hard Attention: A Tale of Two Sides
Selim Jerad
Anej Svete
Jiaoda Li
Ryan Cotterell
49
0
0
18 Mar 2025
Self-Training Elicits Concise Reasoning in Large Language Models
Self-Training Elicits Concise Reasoning in Large Language Models
Tergel Munkhbat
Namgyu Ho
S. Kim
Yongjin Yang
Yujin Kim
Se-Young Yun
ReLM
LRM
48
10
0
27 Feb 2025
Chain-of-Thought in Large Language Models: Decoding, Projection, and
  Activation
Chain-of-Thought in Large Language Models: Decoding, Projection, and Activation
H. Yang
Qianghua Zhao
Lei Li
AI4CE
LRM
61
1
0
05 Dec 2024
An $\mathbf{L^*}$ Algorithm for Deterministic Weighted Regular Languages
An L∗\mathbf{L^*}L∗ Algorithm for Deterministic Weighted Regular Languages
Clemente Pasti
Talu Karagöz
Anej Svete
Franz Nowak
Reda Boumasmoud
Ryan Cotterell
32
0
0
09 Nov 2024
From Sparse Dependence to Sparse Attention: Unveiling How Chain-of-Thought Enhances Transformer Sample Efficiency
From Sparse Dependence to Sparse Attention: Unveiling How Chain-of-Thought Enhances Transformer Sample Efficiency
Kaiyue Wen
Huaqing Zhang
Hongzhou Lin
Jingzhao Zhang
MoE
LRM
42
2
0
07 Oct 2024
Can Transformers Learn $n$-gram Language Models?
Can Transformers Learn nnn-gram Language Models?
Anej Svete
Nadav Borenstein
M. Zhou
Isabelle Augenstein
Ryan Cotterell
19
6
0
03 Oct 2024
Questioning Internal Knowledge Structure of Large Language Models
  Through the Lens of the Olympic Games
Questioning Internal Knowledge Structure of Large Language Models Through the Lens of the Olympic Games
Juhwan Choi
Youngbin Kim
30
0
0
10 Sep 2024
Representing Rule-based Chatbots with Transformers
Representing Rule-based Chatbots with Transformers
Dan Friedman
Abhishek Panigrahi
Danqi Chen
44
1
0
15 Jul 2024
Lower Bounds on the Expressivity of Recurrent Neural Language Models
Lower Bounds on the Expressivity of Recurrent Neural Language Models
Anej Svete
Franz Nowak
Anisha Mohamed Sahabdeen
Ryan Cotterell
19
0
0
29 May 2024
Transformers Can Represent $n$-gram Language Models
Transformers Can Represent nnn-gram Language Models
Anej Svete
Ryan Cotterell
24
17
0
23 Apr 2024
Chain-of-Thought Reasoning Without Prompting
Chain-of-Thought Reasoning Without Prompting
Xuezhi Wang
Denny Zhou
ReLM
LRM
135
97
0
15 Feb 2024
Large Language Models are Zero-Shot Reasoners
Large Language Models are Zero-Shot Reasoners
Takeshi Kojima
S. Gu
Machel Reid
Yutaka Matsuo
Yusuke Iwasawa
ReLM
LRM
291
2,712
0
24 May 2022
1