My research interests are primarily in the field of natural language processing.
In particular, I am interested in expanding the scope of traditional language tasks to be more inclusive of low-resource languages and non-standard dialects.
Papers
-
Stereotype or Personalization? User Identity Biases Chatbot Recommendations
Anjali Kantharuban*, Jeremiah Milbauer*, Emma Strubell, Graham Neubig
Under Review at TACL
-
Pangea: A Fully Open Multilingual Multimodal LLM for 39 Languages
Xiang Yue*, Yueqi Song*, Akari Asai, Seungone Kim, Jean de Dieu Nyandwi, Simran Khanuja, Anjali Kantharuban, Lintang Sutawika, Sathyanarayanan Ramamoorthy, Graham Neubig
Under Review at ICLR, 2025
-
Quantifying the Dialect Gap and its Correlates Across Languages
Anjali Kantharuban, Ivan Vulić, Anna Korhonen
Findings of EMNLP, 2023
-
Counting the Bugs in ChatGPT's Wugs: A Multilingual Investigation into the Morphological Capabilities of a Large Language Model
Leonie Weissweiler, Valentin Hofmann, Anjali Kantharuban, Anna Cai, Ritam Dutt, Amey Hengle, Anubha Kabra, Atharva Kulkarni, Abhishek Vijayakumar, Haofei Yu, Hinrich Schütze, Kemal Oflazer, David R. Mortensen
EMNLP, 2023
-
One-Shot Lexicon Learning for Low-Resource Machine Translation
Anjali Kantharuban, Jacob Andreas
Workshop on Widening Natural Language Processing (WiNLP), EMNLP, 2021
Presentations