Yu Ying Chiu (Kelly Chiu)

MS student at University of Washington, Research Associate at New York University.

icon.jpg

kellycyy [at] uw [dot] edu

I’m a final-year master’s student in Computational Linguistics (NLP) at the University of Washington, advised by Prof. Yejin Choi and Prof. Tim Althoff. I also worked closely with Prof. Shane Steinert-Threlkeld.

I was a ML Alignment & Theory Scholar (MATS), worked with Evan Hubinger from Anthropic on AI values evaluations (LitmusValues).

Currently, I’m a Research Associate at New York University, working with Prof. Sydney Levine from NYU/Google DeepMind and Prof. Mitchell Gordon from MIT/OpenAI. Check out our recent work, MoReBench (a PhD-level, rubric-based moral reasoning benchmark), in collaboration with Scale AI!

My research focuses on the intersection of Human-centered AI and AI Safety. Broadly, I’m intrigued by how humans and AI understand the world, think and make important decisions, and how we can use these insights to better support humans using AI and ensure AI can benefit humanity . My goal is to build and enhance moral, empathetic, and culturally sensitive AI, with a current focus on real-world data collection, human-AI collaboration and value-alignment evaluation.

Always happy to connect with people who are excited to chat or collaborate on AI value alignment and AI safety research!

selected publications

  1. Under review
    MoReBench: Evaluating Procedural and Pluralistic Moral Reasoning in Language Models, More than Outcomes
    Yu Ying Chiu, Michael S. Lee, Rachel Calcott, Brandon Handoko, Paul Font-Reaulx, Paula Rodriguez, Chen Bo Calvin Zhang, Ziwen Han, Udari Madhushani Sehwag, Yash Maurya, Christina Q Knight, Harry R. Lloyd, Florence Bacus, Mantas Mazeika, Bing Liu, Yejin Choi, Mitchell L Gordon, and Sydney Levine
    2025
  2. Under review
    Will AI Tell Lies to Save Sick Children? Litmus-Testing AI Values Prioritization with AIRiskDilemmas
    Yu Ying Chiu, Zhilin Wang, Sharan Maiya, Yejin Choi, Kyle Fish, Sydney Levine, and Evan Hubinger
    2025
  3. ACL 2025
    CulturalBench: a Robust, Diverse and Challenging Benchmark on Measuring the (Lack of) Cultural Knowledge of LLMs Through Human-AI Red-Teaming
    Yu Ying Chiu, Liwei Jiang, Bill Yuchen Lin, Chan Young Park, Shuyue Stella Li, Sahithya Ravi, Mehar Bhatia, Maria Antoniak, Yulia Tsvetkov, Vered Shwartz, and Yejin Choi
    2024
  4. ICLR 2025 (Spotlight)
    DailyDilemmas: Revealing Value Preferences of LLMs with Quandaries of Daily Life
    Yu Ying Chiu, Liwei Jiang, and Yejin Choi
    In International Conference on Learning Representations 2025 (Spotlight), 2024
  5. Under Review
    A Computational Framework for Behavioral Assessment of LLM Therapists
    Yu Ying Chiu*, Ashish Sharma*, Inna Wanyin Lin, and Tim Althoff
    2024
  6. EMNLP 2023 System Demo
    humanoidagent_gif.gif
    Humanoid Agents: Platform for Simulating Human-like Generative Agents
    Zhilin Wang*Yu Ying Chiu*, and Yu Cheung Chiu
    In Empirical Methods in Natural Language Processing 2023 (System Demonstrations), Dec 2023