
Title |
|---|
![]() Private Memorization Editing: Turning Memorization into a Defense to Strengthen Data Privacy in Large Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2025 |
![]() Isotropy, Clusters, and ClassifiersAnnual Meeting of the Association for Computational Linguistics (ACL), 2024 |
![]() The Curious Case of Hallucinatory (Un)answerability: Finding Truths in
the Hidden States of Over-Confident Large Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023 |
![]() Why bother with geometry? On the relevance of linear decompositions of
Transformer embeddingsBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2023 |
![]() Explaining How Transformers Use Context to Build PredictionsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023 |
![]() Token-wise Decomposition of Autoregressive Language Model Hidden States
for Analyzing Model PredictionsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023 |
![]() Dissecting Recall of Factual Associations in Auto-Regressive Language
ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023 |
![]() Jump to Conclusions: Short-Cutting Transformers With Linear
TransformationsInternational Conference on Language Resources and Evaluation (LREC), 2023 |
![]() Understanding Transformer Memorization Recall Through IdiomsConference of the European Chapter of the Association for Computational Linguistics (EACL), 2022 |
![]() Analyzing Transformers in Embedding SpaceAnnual Meeting of the Association for Computational Linguistics (ACL), 2022 |