Karolina Stańczak

Mila, McGill University. Montreal, Canada.

stanczak_photo.jpg

I am a Postdoctoral Researcher at the Quebec AI Institute and McGill University School of Computer Science. I have earned my PhD from the Department of Computer Science at the University of Copenhagen where I was supervised by Isabelle Augenstein and co-supervised by Ryan Cotterell at ETH Zurich. My thesis, titled A Multilingual Perspective on Probing Gender Bias, was awarded the SCIENCE Faculty's PhD Award for advancing innovative techniques to detect gender bias both in natural language and language models.

My research interests encompass interpretability, multilinguality, and safety of large language models, with a focus on developing responsible and transparent AI systems in diverse, multicultural contexts.

Before starting my PhD, I’ve completed the MSc in Statistics at the Humboldt University of Berlin. Before that, I’ve obtained a Bachelor of Science in Economics also at the Humboldt University of Berlin.

Besides, prior to starting my PhD I have worked as a data science consultant for Deloitte Analytics Institute.

You can find me on: Twitter, GitHub, LinkedIn.

news

Nov 11, 2024 I attended EMNLP 2024 and presented there 3 publications: Benchmarking vision language models for cultural understanding, Social bias probing: Fairness benchmarking for language models, and The Causal Influence of Grammatical Gender on Distributional Semantics.
Sep 20, 2024 I was honored to receive the SCIENCE PhD award from the University of Copenhagen.

selected publications

  1. EMNLP 2024
    Benchmarking Vision Language Models for Cultural Understanding
    Shravan Nayak, Kanishk Jain, Rabiul Awal, and 5 more authors
    In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Nov 2024
  2. TACL
    The Causal Influence of Grammatical Gender on Distributional Semantics
    Karolina Stańczak, Kevin Du, Adina Williams, and 2 more authors
    Transactions of the Association for Computational Linguistics, Nov 2024
  3. NAACL 2022
    Same Neurons, Different Languages: Probing Morphosyntax in Multilingual Pre-trained Models
    Karolina Stańczak, Edoardo Ponti, Lucas Torroba Hennigen, and 2 more authors
    In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Jul 2022
  4. EMNLP 2024
    Social Bias Probing: Fairness Benchmarking for Language Models
    Marta Marchiori Manerba*Karolina Stańczak*, Riccardo Guidotti, and 1 more author
    In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, Nov 2024