Debiasing Large Vision-Language Models by Ablating Protected Attribute Representations
Published in NeurIPS 2024 SafeGenAI Workshop, 2024
Supervisors: Andrew C. Eggers, Raymond Duch
Wrote doctoral thesis on applications of RL and NLP in political campaigns:
Teaching: Python, statistics and HPCs to social scientists.
Supervisor: Andrew C. Eggers
Focused on quantitative/statistical methods for causal inference and NLP.
Read Philosophy, Politics and Economics at Merton College, Oxford.
Staff scientist with Multimodal Cognitive AI group under PI Vasudev Lal.
Research Highlights:
Engineering Highlights:
Postdoctoral researcher at Data Driven Social Science Initiative working with Professor Brandon Stewart.
Predoctoral researcher on UKRI-funded MENMOPE project led by Professor Lucy Barnes.
Data scientist/analyst at SBI Japannext, Japan’s foremost PTS.
Experienced with following:
Python
, R
, bash
, SQL
pytorch
, transformers
torch.distributed
, deepspeed
, FSDP
docker
, kubernetes
and SLURM
Published in NeurIPS 2024 SafeGenAI Workshop, 2024
Published in NeurIPS 2024 Creativity and Generative AI, 2024
We use model steering to measure and enhance creative properties of LLMs.
Published in NeurIPS 2024 Tackling Climate Change with Machine Learning, 2024
Published in EMNLP 2024 (Main), 2024
We introduce a three-part framework for constructing persuasive messages.
Published in EMNLP 2024 (Findings), 2024
Including an image in the query to a LLaVA-style VLM significantly increases the likelihood of the model returning an English response, regardless of the language of the query.
Published in CVPR 2024 Multimodal Foundation Models Workshop, 2024
We train a suite of multimodal foundation models (MMFM) using the popular LLaVA framework with the recently released Gemma family of large language models (LLMs).
Published in ACL 2024 (Main), 2024
We evaluate the latent political biases of LLMs, and show that models give substantively different answers when not forced; that answers change depending on how models are forced; and that answers lack paraphrase robustness.
Published in NeurIPS 2023 (Main), 2023
We present a new algorithm for using imperfect annotation surrogates for downstream statistical analyses while guaranteeing statistical properties—like asymptotic unbiasedness and proper uncertainty quantification—which are fundamental to Computational Social Science research.