Publications

(2024). Loss in the Crowd: Hidden Breakthroughs in Language Model Training. ICML Workshop on Mechanistic Interpretability.
(2024). First Tragedy, then Parse: History Repeats Itself in the New Era of Large Language Models. North American Association for Computational Linguistics (NAACL).
(2024). Fast Forwarding Low-Rank Training. Empirical Methods in Natural Language Processing (EMNLP).
(2024). Dynamic Masking Rate Schedules for MLM Pretraining. European Association for Computational Linguistics (EACL).
(2024). ChatGPT Doesn't Trust Chargers Fans: Guardrail Sensitivity in Context. Empirical Methods in Natural Language Processing (EMNLP).
(2024). Causation Does Not Imply Correlation: A Study of Circuit Mechanisms and Model Behaviors. NeurIPS Workshop on Scientific Methods for Understanding Deep Learning.
(2024). Benchmarks as Microscopes: A Call for Model Metrology. Conference on Language Modeling (COLM).
(2024). Attribute Diversity Determines the Systematicity Gap in VQA. Empirical Methods in Natural Language Processing (EMNLP).
(2023). Towards out-of-distribution generalization in large-scale astronomical surveys: robust networks learn similar representations. NeurIPS workshop on Machine Learning and the Physical Sciences.
(2023). State-of-the-art generalisation research in NLP: a taxonomy and review. Nature Machine Intelligence.