default search action
6th BlackboxNLP@EMNLP 2023: Singapore
- Yonatan Belinkov, Sophie Hao, Jaap Jumelet, Najoung Kim, Arya McCarthy, Hosein Mohebbi:
Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, BlackboxNLP@EMNLP 2023, Singapore, December 7, 2023. Association for Computational Linguistics 2023, ISBN 979-8-89176-052-3 - Anthony M. Colas, Jun Araki, Zhengyu Zhou, Bingqing Wang, Zhe Feng:
Knowledge-Grounded Natural Language Recommendation Explanation. 1-15 - Neel Nanda, Andrew Lee, Martin Wattenberg:
Emergent Linear Representations in World Models of Self-Supervised Sequence Models. 16-30 - Chandan Singh, John X. Morris, Jyoti Aneja, Alexander M. Rush, Jianfeng Gao:
Explaining Data Patterns in Natural Language with Language Models. 31-55 - Akshat Gupta:
Probing Quantifier Comprehension in Large Language Models: Another Example of Inverse Scaling. 56-64 - Stefan Arnold, Nils Kemmerzell, Annika Schreiner:
Disentangling the Linguistic Competence of Privacy-Preserving BERT. 65-75 - Antoine Chaffin, Julien Delaunay:
"Honey, Tell Me What's Wrong", Global Explanation of Textual Discriminative Models through Cooperative Generation. 76-88 - Henning Bartsch, Ole Jorgensen, Domenic Rosati, Jason Hoelscher-Obermaier, Jacob Pfau:
Self-Consistency of Large Language Models under Ambiguity. 89-105 - Hao Sun, John Hewitt:
Character-Level Chinese Backpack Language Models. 106-119 - Sunit Bhattacharya, Ondrej Bojar:
Unveiling Multilinguality in Transformer Models: Exploring Language Specificity in Feed-Forward Networks. 120-126 - Timothee Mickus, Raúl Vázquez:
Why Bother with Geometry? On the Relevance of Linear Decompositions of Transformer Embeddings. 127-141 - Dmitry Nikolaev, Sebastian Padó:
Investigating Semantic Subspaces of Transformer Sentence Embeddings through Linear Structural Probing. 142-154 - Juanhe (TJ) Tan:
Causal Abstraction for Chain-of-Thought Reasoning in Arithmetic Word Problems. 155-168 - Natalia Flechas Manrique, Wanqian Bao, Aurélie Herbelot, Uri Hasson:
Enhancing Interpretability Using Human Similarity Judgements to Prune Word Embeddings. 169-179 - Judith Sieker, Sina Zarrieß:
When Your Language Model Cannot Even Do Determiners Right: Probing for Anti-Presuppositions and the Maximize Presupposition! Principle. 180-198 - Jonas Groschwitz:
Introducing VULCAN: A Visualization Tool for Understanding Our Models and Data by Example. 199-211 - David Kletz, Pascal Amsili, Marie Candito:
The Self-Contained Negation Test Set. 212-221 - Yan Cong, Emmanuele Chersoni, Yu-Yin Hsu, Philippe Blache:
Investigating the Effect of Discourse Connectives on Transformer Surprisal: Language Models Understand Connectives, Even So They Are Surprised. 222-232 - Yichu Zhou, Vivek Srikumar:
METAPROBE: A Representation- and Task-Agnostic Probe. 233-249 - Jacob K. Johnson, Ana Marasovic:
How Much Consistency Is Your Accuracy Worth? 250-260 - Tanja Baeumel, Soniya Vijayakumar, Josef van Genabith, Guenter Neumann, Simon Ostermann:
Investigating the Encoding of Words in BERT's Neurons Using Feature Textualization. 261-270 - Shunjie Wang, Shane Steinert-Threlkeld:
Evaluating Transformer's Ability to Learn Mildly Context-Sensitive Languages. 271-283 - Nirmalendu Prakash, Roy Ka-Wei Lee:
Layered Bias: Interpreting Bias in Pretrained Large Language Models. 284-295 - Isabelle Lorge, Janet B. Pierrehumbert:
Not Wacky vs. Definitely Wacky: A Study of Scalar Adverbs in Pretrained Language Models. 296-316 - Jing Huang, Atticus Geiger, Karel D'Oosterlinck, Zhengxuan Wu, Christopher Potts:
Rigorously Assessing Natural Language Explanations of Neurons. 317-331 - Deanna DeCarlo, William Palmer, Michael Wilson, Bob Frank:
NPIs Aren't Exactly Easy: Variation in Licensing across Large Language Models. 332-341 - Mansi Sakarvadia, Aswathy Ajith, Arham Khan, Daniel Grzenda, Nathaniel Hudson, André Bauer, Kyle Chard, Ian T. Foster:
Memory Injections: Correcting Multi-Hop Reasoning Failures During Inference in Transformer-Based Language Models. 342-356 - Aishik Chakraborty, Jackie C. K. Cheung, Timothy J. O'Donnell:
Systematic Generalization by Finetuning? Analyzing Pretrained Language Models Using Constituency Tests. 357-366 - Chenxin Liu, Emmanuele Chersoni:
On Quick Kisses and How to Make Them Count: A Study on Event Construal in Light Verb Constructions with BERT. 367-378 - Abhijith Chintam, Rahel Beloch, Willem H. Zuidema, Michael Hanna, Oskar van der Wal:
Identifying and Adapting Transformer-Components Responsible for Gender Bias in an English Language Model. 379-394
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.