Debiasing Large Vision-Language Models by Ablating Protected Attribute Representations
Published in NeurIPS 2024 SafeGenAI Workshop, 2024
Published in NeurIPS 2024 SafeGenAI Workshop, 2024
Published in NeurIPS 2024 Creativity and Generative AI, 2024
We use model steering to measure and enhance creative properties of LLMs.
Published in NeurIPS 2024 Tackling Climate Change with Machine Learning, 2024
Published in EMNLP 2024 (Main), 2024
We introduce a three-part framework for constructing persuasive messages.
Published in EMNLP 2024 (Findings), 2024
Including an image in the query to a LLaVA-style VLM significantly increases the likelihood of the model returning an English response, regardless of the language of the query.
Published in CVPR 2024 Multimodal Foundation Models Workshop, 2024
We train a suite of multimodal foundation models (MMFM) using the popular LLaVA framework with the recently released Gemma family of large language models (LLMs).
Published in ACL 2024 (Main), 2024
We evaluate the latent political biases of LLMs, and show that models give substantively different answers when not forced; that answers change depending on how models are forced; and that answers lack paraphrase robustness.
Published in NeurIPS 2023 (Main), 2023
We present a new algorithm for using imperfect annotation surrogates for downstream statistical analyses while guaranteeing statistical properties—like asymptotic unbiasedness and proper uncertainty quantification—which are fundamental to Computational Social Science research.