Estimating the Micro-Targeting Effect

Evidence from a Survey Experiment
during the 2020 US Election

Musashi Harukawa

Politics in Progress Colloquium, HT21

Introduction

tl;dr

  • Survey experiment during 2020 US election to estimate effect of micro-targeting.
  • Trained algorithm to target participants with anti-Biden ads run by Trump campaign.
  • Among unaligned respondents who had not pre-voted at time of survey targeting:
    • Increased proportion anti-Biden by 8.7 percentage points.
    • Decreased proportion intending to vote Biden by 7.1 percentage points.

Definitions and Scope

  • Tailoring is constructing a message so that it appeals to a specific audience.
  • Targeting is delivering the message so only the intended audience sees it.
  • Micro- is on the basis of individual characteristics.

Context

  • In 2016 Cambridge Analytica reportedly able to affect election outcomes using individuals’ Facebook data (Simon 2019).
  • Significant media and scholarly attention warning of threats and consequences for society:
    • informational “filter bubbles” threaten civil discourse (e.g. Burkell & Regan 2020).
  • Many of these arguments presume that micro-targeting works:
    • “micro-targeting of voters can pay very handsome electoral dividends for a relatively modest investment” (Krotoszynski Jr. 2020).

Contradicting Evidence

  • Decade of political science research leaves little space for micro-targeting to make a difference.
    • Coppock et al (2020) test 49 advertisements on 34,000 people, find little evidence of heterogeneous effects.
  • Psychology research simulating targeting suggests it should work:
    • Madsen and Pilditch (2018) use ABM.
    • Zarouali et al (2020) use experiment (N=158).

Gaps and Challenges

  • Strategies and algorithms proprietary (Edelson et al 2019).
  • Data difficult to obtain (Liberini et al 2020).

Research Question

Does micro-targeting work?

or:

Is it possible to improve the effectiveness of a campaign by optimally allocating advertisements on the basis of individual traits?

Research Design

Case Selection

  • US 2020 Presidential Election
  • Participants US citizens, resident in US, of voting age.
  • Payment and recruitment via Prolific.
  • Redirected to external website https://survey.polinfo.org

Design Summarized

  • Five advertisements.
  • First stage (N=1,500), respondents shown random ad.
  • This data used to train targeting algorithm.
  • Second stage (N=900), respondents shown optimal ad.
  • Difference between Stage 1 and 2 is treatment effect of targeting.

Advertisements

Title Description
“They Mock Us” In-group: Clinton and Biden mocking
“Why did Biden let him do it?” Hunter Biden’s ostensible corruption
“Biden will come for your guns” 2A; Biden will steal guns
“Insult” Biden: Black Trump supporters not Black
“Real Leadership” Obama/Biden caused wars, neglected veterans

Stage 1

For individual \(i \in \{1,...,1500\}\):

  • Pre-treatment covariates \(\mathbf{X}_i\).
  • Shown \(d_a\), where \(a \sim \mathcal{U}\{1, 5\}\).
  • Post-treatment outcome(s) \(Y_i\)
    • Biden and Trump favorability (1-5)
    • Voting preference

Predicting Optimal Advertisement

Results of stage 1 used to learn outcome as function of pre-treatment traits and advertisement:

\[ Y_i = f(X_i, d_a) \]

Thus for any person, I can predict their hypothetical outcome under each of the five advertisements:

\[ \hat{f}(X_i, d_1) = \hat{Y}(d_1) \] \[ \{\hat{Y}(d_1), \hat{Y}(d_2), ... \hat{Y}(d_5)\} \]

Five Candidate Algorithms

Chosen for speed and ability to learn highly conditional relationships:

  • Random Forest (RF)
  • AdaBoost
  • Gradient Boosted Decision Trees (GBDT)
  • Multi-Layer Perceptron Regressor (MLPR)
  • Support Vector Machine (SVM)

Choosing the Best Model

  • Models trained/tested with 30-fold cross-validation
  • Compared on RMSE, max error and prediction time
  • RF and AdaBoost generally better, RF weakly better than AdaBoost and less likely to give ties
  • Fitted model uploaded to web server

Stage 2

  • Same pre-treatment questions \(\mathbf{X}_i\).
  • Predictive model chooses optimal advertisement as a function of \(\mathbf{X}_i\):

\[ d^*(X_i): opt_a \; f(X_i, d_{i, a}) \]

Average Targeting Effect

Compare average outcome between randomly assigned stage 1 and optimally assigned stage 2:

\[ ATE = \mathbb{E}_i[Y_i(d^*(X_i))] - \mathbb{E}_a[\mathbb{E}_i[Y_i(d_{i, a})]] \]

Hypotheses

Three outcomes of interest:

  • Hypothesis 1 (Micro-targeting Affects Favorability): \(\mathbb{E}_i[y_{i, Biden}(d^*)] < \mathbb{E}_a[\mathbb{E}_i[y_{i, Biden}(d_{i, a})]]\)
  • Hypothesis 2 (Micro-targeting Affects Voting Preference): \(\mathbb{E}_i[v_{i, Biden}(d^*)] < \mathbb{E}_a[\mathbb{E}_i[v_{i, Biden}(d_{i, a})]]\)
  • Hypothesis 3 (Micro-targeting Affects Turnout): \(\mathbb{E}_i[u_{i}(d^*)] < \mathbb{E}_a[\mathbb{E}_i[u_{i}(d_{i, a})]]\)

CATEs of Interest

Account for conditional effect of two pre-treatment covariates:

  • Partisan self-identification
  • Early voting

Results

Stage 1 - Treatment Assignments

Stage 1 - Feature Importances

Stage 1 - Predicted Outcome as Function of Partisanship

Stage 2

Robustness

  • Pre-treatment covariate balance check.
  • Multiple comparisons correction (Holm, Benjamini-Hochberg).
  • Variety of operationalizations of outcome (linear, binary, ordered categorical)
  • Pre-experimental power check using Coppock et al (2020) data (along with permutation test for bias in mechanism).

Discussion

Key Results

  • Among unaligned respondents who had not pre-voted at time of the survey, targeting:
    • Increased proportion anti-Biden by 8.7 percentage points.
    • Decreased proportion intending to vote Biden by 7.1 percentage points.
  • These margins are greater than those seen in many crucial districts.

Normative Aspects

  • Data and algorithms to influence the outcome of elections?
    • Whose preferences are being represented?
    • Creates incentives to harvest voter data (Krotoszynski Jr. 2020)
  • When is it permissible to target political advertisements?
    • Is negative advertising the problem?
    • Issues of consent and privacy.

Ethical Conisderations

  • CUREC approved
  • Participant consent obtained
  • Extensive end-of-survey debrief detailing purpose and potential manipulation

Limitations

  • Survey Response vs. Vote Choice
  • Convenience Sample
  • Psychometric Profiling vs Data-Driven Optimization
    • Relatedly: Mechanisms?
  • Reference Category?

What’s Next

  • Follow up experiment planned, Germany?
    • Additional treatments to test effect of informing participants of targeting.
  • Further exploration of normative aspects
// reveal.js plugins