Urja Khurana

I am a fourth-year PhD student at the Computational Linguistics and Text Mining Lab (CLTL) at the Vrije Universiteit Amsterdam. I am advised by Antske Fokkens (Vrije Universiteit Amsterdam) and Eric Nalisnick (Johns Hopkins University).

I am passionate about understanding what knowledge a language model has captured and whether it can reliably apply this knowledge across diverse (unseen) contexts; what will its impact be in the real world? I am particularly interested in analyzing and developing tools to ensure responsible deployment, with a big focus on robustness for safety-critical applications, e.g. hate speech detection.

My work includes evaluating the generalization of model capabilities to unseen data, for instance through the lens of model averaging and consistency. I have also proposed a method to calibrate language models according to human subjectivity. Collaborating in an interdisciplinary environment with a social scientist and law expert, I characterized the subjective aspects of hate speech and its impact on real-world deployment. Building on these insights, I developed a framework to evaluate if a hate speech detection model’s behavior aligns with the type of hate speech it is intended to address.

I earned my BSc and MSc degrees in Artificial Intelligence from the University of Amsterdam.

If you want to (briefly) know more about my research interests, see the research page.

selected publications.

  1. DefVerify: Do Hate Speech Models Reflect Their Dataset's Definition?
    Khurana, Urja, Nalisnick, Eric, and Fokkens, Antske
    To appear at COLING 2025
  2. Crowd-Calibrator: Can Annotator Disagreement Inform Calibration in Subjective Tasks?
    Khurana, Urja, Nalisnick, Eric, Fokkens, Antske, and Swayamdipta, Swabha
    In First Conference on Language Modeling 2024
  3. Hate Speech Criteria: A Modular Approach to Task-Specific Hate Speech Definitions
    Khurana, Urja, Vermeulen, Ivar, Nalisnick, Eric, Van Noorloos, Marloes, and Fokkens, Antske
    In Proceedings of the Sixth Workshop on Online Abuse and Harms (WOAH) 2022
  4. How Emotionally Stable is {ALBERT}? Testing Robustness with Stochastic Weight Averaging on a Sentiment Analysis Task
    Khurana, Urja, Nalisnick, Eric, and Fokkens, Antske
    In Proceedings of the 2nd Workshop on Evaluation and Comparison of NLP Systems 2021