ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2004.01254
  4. Cited By
Under the Hood of Neural Networks: Characterizing Learned
  Representations by Functional Neuron Populations and Network Ablations
v1v2 (latest)

Under the Hood of Neural Networks: Characterizing Learned Representations by Functional Neuron Populations and Network Ablations

2 April 2020
Richard Meyes
Constantin Waubert de Puiseau
Andres Felipe Posada-Moreno
Tobias Meisen
    AI4CE
ArXiv (abs)PDFHTML

Papers citing "Under the Hood of Neural Networks: Characterizing Learned Representations by Functional Neuron Populations and Network Ablations"

15 / 15 papers shown
Structural Inference: Interpreting Small Language Models with Susceptibilities
Structural Inference: Interpreting Small Language Models with Susceptibilities
Garrett Baker
George Wang
Jesse Hoogland
Daniel Murfet
AAML
502
1
0
25 Apr 2025
Differentiation and Specialization of Attention Heads via the Refined
  Local Learning Coefficient
Differentiation and Specialization of Attention Heads via the Refined Local Learning CoefficientInternational Conference on Learning Representations (ICLR), 2024
George Wang
Jesse Hoogland
Stan van Wingerden
Zach Furman
Daniel Murfet
OffRL
285
28
0
03 Oct 2024
Modularity in Transformers: Investigating Neuron Separability &
  Specialization
Modularity in Transformers: Investigating Neuron Separability & Specialization
Nicholas Pochinkov
Thomas Jones
Mohammed Rashidur Rahman
206
0
0
30 Aug 2024
Modular Deep Learning
Modular Deep Learning
Jonas Pfeiffer
Sebastian Ruder
Ivan Vulić
Edoardo Ponti
MoMeOOD
479
109
0
22 Feb 2023
Much Easier Said Than Done: Falsifying the Causal Relevance of Linear
  Decoding Methods
Much Easier Said Than Done: Falsifying the Causal Relevance of Linear Decoding Methods
L. Hayne
Abhijit Suresh
Hunar Jain
Rahul Kumar
R. M. Carter
FAtt
235
1
0
08 Nov 2022
Toward Transparent AI: A Survey on Interpreting the Inner Structures of
  Deep Neural Networks
Toward Transparent AI: A Survey on Interpreting the Inner Structures of Deep Neural Networks
Tilman Raukur
A. Ho
Stephen Casper
Dylan Hadfield-Menell
AAMLAI4CE
861
182
0
27 Jul 2022
Discovering Salient Neurons in Deep NLP Models
Discovering Salient Neurons in Deep NLP ModelsJournal of machine learning research (JMLR), 2022
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
KELMMILM
360
20
0
27 Jun 2022
Future Artificial Intelligence tools and perspectives in medicine
Future Artificial Intelligence tools and perspectives in medicineCurrent Opinion in Urology (Curr Opin Urol), 2021
Ahmad Chaddad
Y. Katib
Lama Hassan
297
10
0
04 Jun 2022
Neuron-level Interpretation of Deep NLP Models: A Survey
Neuron-level Interpretation of Deep NLP Models: A SurveyTransactions of the Association for Computational Linguistics (TACL), 2021
Hassan Sajjad
Nadir Durrani
Fahim Dalvi
MILMAI4CE
435
97
0
30 Aug 2021
How to Certify Machine Learning Based Safety-critical Systems? A
  Systematic Literature Review
How to Certify Machine Learning Based Safety-critical Systems? A Systematic Literature ReviewInternational Conference on Automated Software Engineering (ASE), 2021
Florian Tambon
Gabriel Laberge
Le An
Amin Nikanjam
Paulina Stevia Nouwou Mindom
Y. Pequignot
Foutse Khomh
G. Antoniol
E. Merlo
François Laviolette
590
87
0
26 Jul 2021
Dynamics of specialization in neural modules under resource constraints
Dynamics of specialization in neural modules under resource constraintsNature Communications (Nat Commun), 2021
Gabriel Béna
Dan F. M. Goodman
378
9
0
04 Jun 2021
R2U3D: Recurrent Residual 3D U-Net for Lung Segmentation
R2U3D: Recurrent Residual 3D U-Net for Lung SegmentationIEEE Access (IEEE Access), 2021
Dhaval Kadia
Md. Zahangir Alom
Ranga Burada
Tam V. Nguyen
V. Asari
3DPCSSeg
246
16
0
05 May 2021
Contrastive Explanations for Model Interpretability
Contrastive Explanations for Model InterpretabilityConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Alon Jacovi
Swabha Swayamdipta
Shauli Ravfogel
Yanai Elazar
Yejin Choi
Yoav Goldberg
525
115
0
02 Mar 2021
Explainable deep learning models in medical image analysis
Explainable deep learning models in medical image analysisJournal of Imaging (JI), 2020
Amitojdeep Singh
S. Sengupta
Vasudevan Lakshminarayanan
XAI
413
610
0
28 May 2020
A Survey on Explainable Artificial Intelligence (XAI): Towards Medical
  XAI
A Survey on Explainable Artificial Intelligence (XAI): Towards Medical XAIIEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2019
Erico Tjoa
Cuntai Guan
XAI
785
1,897
0
17 Jul 2019
1
Page 1 of 1