Source Themes

SHADES: Towards a multilingual assessment of stereotypes in large language models (2025, NAACL)

Click the Cite button above to demo the feature to enable visitors to import publication metadata into their reference management software. # Click the Slides button above to demo Academic’s Markdown slides feature.

The Root Shapes the Fruit: On the Persistence of Gender-Exclusive Harms in Aligned Language Models (2025, ACM FAccT)

Click the Cite button above to demo the feature to enable visitors to import publication metadata into their reference management software. # Click the Slides button above to demo Academic’s Markdown slides feature.

Tokenization Matters: Navigating Data-Scarce Tokenization for Gender Inclusive Language Technologies (2024, NAACL Findings)

Click the Cite button above to demo the feature to enable visitors to import publication metadata into their reference management software. # Click the Slides button above to demo Academic’s Markdown slides feature.

β€œI’m fully who I am”: Towards Centering Transgender and Non-Binary Voices to Measure Biases in Open Language Generation (ACM FAccT, 2023)

Transgender and non-binary (TGNB) individuals disproportionately experience discrimination and exclusion from daily life. Given the recent popularity and adoption of language generation technologies, the potential to further marginalize this …

Queer In AI: A Case Study in Community-Led Participatory AI (ACM FAcct Best Paper Award, 2023)

Queerness and queer people face an uncertain future in the face of ever more widely deployed and invasive artificial intelligence (AI). These technologies have caused numerous harms to queer people, including privacy violations, censoring and …

Bound by the Bounty: Collaboratively Shaping Evaluation Processes for Queer AI Harms

Bias evaluation benchmarks and dataset and model documentation have emerged as central processes for assessing the biases and harms of artificial intelligence (AI) systems. However, these auditing processes have been criticized for their failure to …

Factoring the Matrix of Domination: A Critical Review and Reimagination of Intersectionality in AI Fairness

Intersectionality is a critical framework that, through inquiry and praxis, allows us to examine how social inequalities persist through domains of structure and discipline. Given AI fairness' raison d'Γͺtre of 'fairness', we argue that adopting …

Should they? Mobile Biometrics and Technopolicy meet Queer Community Considerations

Smartphones are integral to our daily lives and activities, providing us with basic functions like texting and phone calls to more complex motion-based functionalities like navigation, mobile gaming, and fitness-tracking. To facilitate these …

Improving Adversarial Robustness to Sensitivity and Invariance Attacks with Deep Metric Learning

Intentionally crafted adversarial samples have effectively exploited weaknesses in deep neural networks. A standard method in adversarial robustness assumes a framework to defend against samples crafted by minimally perturbing a sample such that its …

Auditing Algorithmic Fairness in Machine Learning for Health with Severity-Based LOGAN

Auditing machine learning-based (ML) healthcare tools for bias is critical to preventing patient harm, especially in communities that disproportionately face health inequities. General frameworks are becoming increasingly available to measure ML …