Drawing from Human-Computer Interaction (HCI), Machine Learning (ML), Artificial Intelligence (AI), Organizational Science, Design, Psychology, Crowdsourcing, and Social Computing, my work aims at supporting on-the-ground responsible AI (RAI) practice. I achieve this through an iterative process of empirical understanding and system development, facilitating the meaningful communication and collabroation between AI practitioners and diverse end-users. I mostly publish at HCI and responsible AI (RAI) venues such as CHI, CSCW, FAccT, AIES, and HCOMP.
WeAudit: Scaffolding User Auditors and AI Practitioners in Auditing Generative AI.
arXiv ●
PRE-PRINT
* Invited talks at Apple Human-Centered Machine Learning team; Big Design Seminar at Zhejiang University.
Featured as educational materials for more than 600 students across 10 classes at CMU.
Supporting Industry Computing Researchers in Assessing, Articulating, and Addressing the Potential Negative Societal Impact of Their Work
CSCW 2025 ●
PAPER ●
TEMPLATE ●
PROJECT PAGE (coming soon)
Human-centered Evaluation and Auditing of Language Models
CHI 2025 Workshop Proposal ●
WEBSITE
Red-Teaming for Generative AI: Silver Bullet or Security Theater?
AIES 2024 ●
Best Paper Award ●
PAPER
Investigating What Factors Influence Users’ Detection of Harmful Algorithmic Bias and Discrimination.
HCOMP 2024 ●
Best Paper Award ●
PAPER
Responsible Crowdsourcing for Responsible Generative AI: Engaging Crowds in AI Auditing and Evaluation
HCOMP 2024 Workshop Proposal ●
PROPOSAL ●
WEBSITE
MIRAGE: Multi-model Interface for Reviewing and Auditing Generative Text-to-Image AI.
HCOMP 2024 DEMO ●
PAPER ●
DEMO
Designing a Crowdsourcing Pipeline to Verify Reports from User AI Audits
HCOMP 2024 WiP ●
PAPER
Human-centered Evaluation and Auditing of Language Models
CHI 2024 Workshop Proposal ●
PROPOSAL ●
WEBSITE
Collaboratively Designing and Evaluating Responsible AI Interventions
CSCW 2024 Workshop Proposal ●
PROPOSAL ●
WEBSITE
Troubling Taxonomies in GenAI Evaluation.
NeurIPS 2024 Workshop on Evaluating Evaluations ●
PAPER
Carefully Unmaking the “Marginalized User”: A Diffractive Analysis of a Gay Online Community.
Transactions on Computer-Human Interaction ●
PAPER
Understanding Practices, Challenges, and Opportunities for User-Engaged Algorithm Auditing in Industry Practice
CHI 2023 ●
PAPER ●
VIDEO
* Invited talks at Faculty of EEMCS and Industrial Design Engineering at TU Delft, Algorithmic Fairness and Opacity (AFOG) Group at UC Berkeley, Tecent Shanghai, Google Shanghai, Salesforce, and Capital One.
Investigating Practices and Opportunities for Cross-functional Collaboration around AI Fairness in Industry Practice
FAccT 2023 ●
PAPER ●
VIDEO
Supporting User Engagement in Testing, Auditing, and Contesting AI
CSCW 2023 Workshop Proposal ●
PROPOSAL ●
WEBSITE
Towards" Anytime, Anywhere" Community Learning and Engagement around the Design of Public Sector AI
CHI 2023 Workshop on AI Literacy ●
PAPER
Exploring How Machine Learning Practitioners (Try To) Use Fairness Toolkits
FAccT 2022 ●
PAPER ●
VIDEO
* Invited talks at FairLearn developer teams at Microsoft Research, AIF360 developer monthly meeting at IBM, People+AI Research team at Google, and AI Governance group at PwC.
The model card authoring toolkit: Toward community-centered, deliberation-driven AI design
FAccT 2022 ●
PAPER ●
VIDEO
Opportunities for human-centered evaluation of machine translation systems
NAACL 2022 ●
PAPER
Beyond General Purpose Machine Translation: The Need for Context-specific Empirical Research to Design for Appropriate User Trust
CHI 2022 Workshop on Trust and Reliance in AI-Human Teaming ●
PAPER
Value cards: An educational toolkit for teaching social impacts of machine learning through deliberation
FAccT 2021 ●
PAPER ●
VIDEO
Robust Laughter Detection in Noisy Environments
Interspeech 2021 ●
PAPER ●
GITHUB
Three Directions for the Design of Human-Centered Machine Translation
ACL HCI + NLP Workshop ●
PAPER ●
Personal laughter archives: reflection through visualization and interaction
UbiComp Work in Progress ●
PAPER