Bago, Bence, Leah R. Rosenzweig, Adam J. Berinsky, and David G. Rand. 2022. “Emotion May Predict Susceptibility to Fake News but Emotion Regulation Does Not Seem to Help.”Cognition and Emotion, June, 1–15. https://doi.org/10.1080/02699931.2022.2090318.
Intervention
Code
intervention_info <-tibble(intervention_description ='Only studies 2, 3 and 4 tested an intervention. The intervention aimed at emotion regulation. In study 2, participants were asked to apply either emotional suppression (`intervention_label` = `emotion_suppression`) or emotion reappraisal techniques (`intervention_label` = `emotion_reappraisal`) when considering the veracity of several headlines. In the emotion reappraisal condition, participants read: "As you view and read the headlines, please try to adopt a detached and unemotional attitude. Please try to think about what you are reading objectively. Read all of the headlines carefully, but please try to think about what you are seeing in such a way that you feel less emotion." In the emotion suppression condition, they read: "As you view and read the headlines, if you have any feelings, please try your best not to let those feelings show. Read all of the headlines carefully, but try to behave so that someone watching you would not know that you are feeling anything at all." In studies 3 and 4 of the paper, the authors only tested the emotion suppression treatment.',intervention_selection ="suppression",intervention_selection_description ='Since we can only include one treatment per control condition in the meta-analysis, we remove participants in the emotion reappraisal condition in study 2. This allows us to compare across their studies.',control_format ="picture, lede, source",originally_identified_treatment_effect =FALSE)# displayshow_conditions(intervention_info)
intervention_description
intervention_selection_description
Only studies 2, 3 and 4 tested an intervention. The intervention aimed at emotion regulation. In study 2, participants were asked to apply either emotional suppression (`intervention_label` = `emotion_suppression`) or emotion reappraisal techniques (`intervention_label` = `emotion_reappraisal`) when considering the veracity of several headlines. In the emotion reappraisal condition, participants read: "As you view and read the headlines, please try to adopt a detached and unemotional attitude. Please try to think about what you are reading objectively. Read all of the headlines carefully, but please try to think about what you are seeing in such a way that you feel less emotion." In the emotion suppression condition, they read: "As you view and read the headlines, if you have any feelings, please try your best not to let those feelings show. Read all of the headlines carefully, but try to behave so that someone watching you would not know that you are feeling anything at all." In studies 3 and 4 of the paper, the authors only tested the emotion suppression treatment.
Since we can only include one treatment per control condition in the meta-analysis, we remove participants in the emotion reappraisal condition in study 2. This allows us to compare across their studies.
Notes
The authors did not find evidence that emotion regulation helped people distinguish false from true news headlines, in any of the studies.
The authors sampled from a pool of 24 news items–a single participant only rated 16 items from this pool
“We used a pool of 24 items, taken from Pennycook and Rand (2019), half of which were real (true) and the other fake (false). Moreover, half of the items were Republican-consistent and the other half were Democrat-consistent items, based on a pre-test. Participants were presented with 16 randomly selected headlines altogether; 4 from each category (i.e. Republican-consistent fake, Republican-consistent real, Democrat-consistent fake, Democrat-consistent real).”
Reference:
Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50. https://doi.org/10.1016/j.cognition.2018.06.011
For study 4, they used different headlines, but again a pool of 24 with a random sample of 16 per participant.
“We used newer headlines taken from Pennycook et al., (2021a).”
Reference:
Pennycook, G., Binnendyk, J., Newton, C., & Rand, D. (2021a). A practical guide to doing behavioural research on fake news and misinformation Collabra: Psychology, 7(1), 25293. https://doi.org/10.1525/collabra.25293
Data Cleaning
Study 2
Read data.
Code
d2 <-read_csv("bago_2022-study_2.csv")
Rows: 16112 Columns: 15
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr (3): reality, condition, concordancy
dbl (11): ID, accuracy, share, click, seen, Gender, libcons, DemRep_C, age, ...
lgl (1): Gender_TEXT
ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.
Code
head(d2)
# A tibble: 6 × 15
ID reality accuracy share click seen Gender Gender_TEXT libcons DemRep_C
<dbl> <chr> <dbl> <dbl> <dbl> <dbl> <dbl> <lgl> <dbl> <dbl>
1 1 real 0 0 0 0 1 NA 1 3
2 1 fake 1 0 0 0 1 NA 1 3
3 1 real 1 0 0 0 1 NA 1 3
4 1 fake 1 0 0 0 1 NA 1 3
5 1 real 1 0 0 0 1 NA 1 3
6 1 real 1 0 0 1 1 NA 1 3
# ℹ 5 more variables: condition <chr>, age <dbl>, concordancy <chr>,
# perceived_accu <dbl>, item <dbl>
veracity
Code
# check levels of accuracy table(d2$reality, useNA ="always")
d2 <- d2 |>mutate(news_id = item, recycled_news =TRUE, recycled_news_reference ="Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50. https://doi.org/10.1016/j.cognition.2018.06.011")
# check levels of concordance table(d2$concordancy, useNA ="always")
Concordant Discordant <NA>
8008 8008 96
From the legend that the authors provide, we know that DemRep_C describes the partisan identity (“Which of the following best describes your political preference? 1: Strong Democrat 2: Democrat 3: Lean Democrat 4: Lean Republican 5: Republican 6: Strong Republican”). We will collapse this variable into a binary one.
# check levels of condition levels(as.factor(d2$condition))
[1] "Control" "Reappraisal" "Suppression"
Code
d2 <- d2 |># only assign a label to interventions conditions, not control conditionsmutate(intervention_label =ifelse(condition =="Control", NA, tolower(condition)),condition =ifelse(condition =="Control", "control", "treatment") )
age
Code
# check levels of agetable(d2$age, useNA ="always") # 98 is likely code for NA
There is no date variable, and in the paper, data collection date is not recorded for Study 2. However, it says that Study 3 has been conducted in September 2020. We therefor use 2020
Rows: 48016 Columns: 12
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr (5): reality, Gender_TEXT, condition, consistent, concordancy
dbl (7): ID, Gender, libcons, DemRep_C, Age, perceived_accu, item
ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.
d3 <- d3 |>mutate(subject_id = ID, news_id = item, recycled_news =TRUE, recycled_news_reference ="Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50. https://doi.org/10.1016/j.cognition.2018.06.011",veracity =ifelse(reality =="fake", "false", "true"),intervention_label =ifelse(condition =="Control", NA, tolower(condition)),condition =ifelse(condition =="Control", "control", "treatment"), accuracy_raw = perceived_accu,scale ="binary",concordance =tolower(concordancy),partisan_identity =ifelse(DemRep_C <=3, "democrat", "republican"),news_slant =case_when(partisan_identity =="democrat"& concordance =="concordant"~"democrat", partisan_identity =="republican"& concordance =="concordant"~"republican", partisan_identity =="democrat"& concordance =="discordant"~"republican", partisan_identity =="republican"& concordance =="discordant"~"democrat", TRUE~NA_character_ ),experiment_id =3,age = Age, year =2020)
Study 4
We proceed as for the previous studies. Note, however, that we add different news source
Code
d4 <-read_csv("bago_2022-study_4.csv")
Rows: 48336 Columns: 14
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr (6): reality, screener1, Gender_TEXT, condition, consistent, concordancy
dbl (8): ID, sceener2, Gender, libcons, DemRep_C, Age, perceived_accu, item
ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.
d4 <- d4 |>mutate(subject_id = ID, news_id = item, recycled_news =TRUE, recycled_news_reference ="Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50. https://doi.org/10.1016/j.cognition.2018.06.011",veracity =ifelse(reality =="fake", "false", "true"),intervention_label =ifelse(condition =="Control", NA, tolower(condition)),condition =ifelse(condition =="Control", "control", "treatment"), accuracy_raw = perceived_accu,concordance =tolower(concordancy),partisan_identity =ifelse(DemRep_C <=3, "democrat", "republican"),news_slant =case_when(partisan_identity =="democrat"& concordance =="concordant"~"democrat", partisan_identity =="republican"& concordance =="concordant"~"republican", partisan_identity =="democrat"& concordance =="discordant"~"republican", partisan_identity =="republican"& concordance =="discordant"~"democrat", TRUE~NA_character_ ),experiment_id =4,age = Age, # again, no direct info on year, we impute year of data collection reported for Study 3year =2020)
Combine and add identifiers (country, paper_id) and news_selection
In study 2 and 3, the same news pool has been used. In study 4, a different pool has been used. We take this into account when merging the different studies.
---title: Emotion May Predict Susceptibility to Fake News but Emotion Regulation Does Not Seem to Help.date: "2022"author: - Bago, Bencecategories: - emotion suppressionbibliography: ../../../references.bibnocite: | @bagoEmotionMayPredict2022draft: false ---```{r}#| label: setup#| include: falselibrary(tidyverse)library(kableExtra)# load functionssource("../../../R/custom_functions.R")# load target variablessource("../../../R/variables.R")```## Reference::: {#refs}:::## Intervention```{r}intervention_info <-tibble(intervention_description ='Only studies 2, 3 and 4 tested an intervention. The intervention aimed at emotion regulation. In study 2, participants were asked to apply either emotional suppression (`intervention_label` = `emotion_suppression`) or emotion reappraisal techniques (`intervention_label` = `emotion_reappraisal`) when considering the veracity of several headlines. In the emotion reappraisal condition, participants read: "As you view and read the headlines, please try to adopt a detached and unemotional attitude. Please try to think about what you are reading objectively. Read all of the headlines carefully, but please try to think about what you are seeing in such a way that you feel less emotion." In the emotion suppression condition, they read: "As you view and read the headlines, if you have any feelings, please try your best not to let those feelings show. Read all of the headlines carefully, but try to behave so that someone watching you would not know that you are feeling anything at all." In studies 3 and 4 of the paper, the authors only tested the emotion suppression treatment.',intervention_selection ="suppression",intervention_selection_description ='Since we can only include one treatment per control condition in the meta-analysis, we remove participants in the emotion reappraisal condition in study 2. This allows us to compare across their studies.',control_format ="picture, lede, source",originally_identified_treatment_effect =FALSE)# displayshow_conditions(intervention_info)```### NotesThe authors did not find evidence that emotion regulation helped people distinguish false from true news headlines, in any of the studies.The authors sampled from a pool of 24 news items--a single participant only rated 16 items from this pool>"We used a pool of 24 items, taken from Pennycook and Rand (2019), half of which were real (true) and the other fake (false). Moreover, half of the items were Republican-consistent and the other half were Democrat-consistent items, based on a pre-test. Participants were presented with 16 randomly selected headlines altogether; 4 from each category (i.e. Republican-consistent fake, Republican-consistent real, Democrat-consistent fake, Democrat-consistent real)."Reference: >Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50. https://doi.org/10.1016/j.cognition.2018.06.011For study 4, they used different headlines, but again a pool of 24 with a random sample of 16 per participant.>"We used newer headlines taken from Pennycook et al., (2021a)."Reference: >Pennycook, G., Binnendyk, J., Newton, C., & Rand, D. (2021a). A practical guide to doing behavioural research on fake news and misinformation Collabra: Psychology, 7(1), 25293. https://doi.org/10.1525/collabra.25293## Data Cleaning### Study 2Read data.```{r}d2 <-read_csv("bago_2022-study_2.csv")head(d2)```#### `veracity````{r}# check levels of accuracy table(d2$reality, useNA ="always")``````{r}d2 <- d2 |>mutate(veracity =ifelse(reality =="fake", "false", "true"))```#### `accuracy_raw` and `scale````{r}# check levels of accuracy table(d2$perceived_accu, useNA ="always")``````{r}d2 <- d2 |>mutate(accuracy_raw = perceived_accu, scale ="binary")```#### `news_id`, `recycled_news`, `recycled_news_reference````{r}d2 |>group_by(item) |>count()```We rename the variable```{r}d2 <- d2 |>mutate(news_id = item, recycled_news =TRUE, recycled_news_reference ="Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50. https://doi.org/10.1016/j.cognition.2018.06.011")```#### Concordance (`concordance`, `partisan_identity`, `news_slant`)A concordance variable is already present.```{r}# check levels of concordance table(d2$concordancy, useNA ="always")```From the legend that the authors provide, we know that `DemRep_C` describes the partisan identity ("Which of the following best describes your political preference? 1: Strong Democrat 2: Democrat 3: Lean Democrat 4: Lean Republican 5: Republican 6: Strong Republican"). We will collapse this variable into a binary one. We alse add a news slant variable. ```{r}d2 <- d2 |>mutate(concordance =tolower(concordancy),partisan_identity =ifelse(DemRep_C <=3, "democrat", "republican"),news_slant =case_when(partisan_identity =="democrat"& concordance =="concordant"~"democrat", partisan_identity =="republican"& concordance =="concordant"~"republican", partisan_identity =="democrat"& concordance =="discordant"~"republican", partisan_identity =="republican"& concordance =="discordant"~"democrat", TRUE~NA_character_ ) )```#### Conditions (`intervention_label`, `condition`)```{r}# check levels of condition levels(as.factor(d2$condition))``````{r}d2 <- d2 |># only assign a label to interventions conditions, not control conditionsmutate(intervention_label =ifelse(condition =="Control", NA, tolower(condition)),condition =ifelse(condition =="Control", "control", "treatment") )```#### `age````{r}# check levels of agetable(d2$age, useNA ="always") # 98 is likely code for NA``````{r}d2 <- d2 |>mutate(age =ifelse(age ==98, NA, age))```#### `year`There is no date variable, and in the paper, data collection date is not recorded for Study 2. However, it says that Study 3 has been conducted in September 2020. We therefor use 2020```{r}d2 <- d2 |>mutate(year =2020)```#### Identifiers (`subject_id`, `experiment_id`) ```{r}d2 <- d2 |>mutate(subject_id = ID, experiment_id =2)```### Study 3We proceed as for the previous study.```{r}d3 <-read_csv("bago_2022-study_3.csv")names(d3)# check levels of condition variablelevels(as.factor(d3$condition))# check levels of agetable(d3$Age, useNA ="always") # 90 seems old, but NA's are coded explictly as suchd3 <- d3 |>mutate(subject_id = ID, news_id = item, recycled_news =TRUE, recycled_news_reference ="Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50. https://doi.org/10.1016/j.cognition.2018.06.011",veracity =ifelse(reality =="fake", "false", "true"),intervention_label =ifelse(condition =="Control", NA, tolower(condition)),condition =ifelse(condition =="Control", "control", "treatment"), accuracy_raw = perceived_accu,scale ="binary",concordance =tolower(concordancy),partisan_identity =ifelse(DemRep_C <=3, "democrat", "republican"),news_slant =case_when(partisan_identity =="democrat"& concordance =="concordant"~"democrat", partisan_identity =="republican"& concordance =="concordant"~"republican", partisan_identity =="democrat"& concordance =="discordant"~"republican", partisan_identity =="republican"& concordance =="discordant"~"democrat", TRUE~NA_character_ ),experiment_id =3,age = Age, year =2020) ```### Study 4We proceed as for the previous studies. Note, however, that we add different news source```{r}d4 <-read_csv("bago_2022-study_4.csv")head(d4)# check levels of condition variablelevels(as.factor(d4$condition))# check levels of agetable(d4$Age, useNA ="always") # again, 90 and 98 seem old, but NA's are coded explictly as suchd4 <- d4 |>mutate(subject_id = ID, news_id = item, recycled_news =TRUE, recycled_news_reference ="Pennycook, G., & Rand, D. G. (2019). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 39–50. https://doi.org/10.1016/j.cognition.2018.06.011",veracity =ifelse(reality =="fake", "false", "true"),intervention_label =ifelse(condition =="Control", NA, tolower(condition)),condition =ifelse(condition =="Control", "control", "treatment"), accuracy_raw = perceived_accu,concordance =tolower(concordancy),partisan_identity =ifelse(DemRep_C <=3, "democrat", "republican"),news_slant =case_when(partisan_identity =="democrat"& concordance =="concordant"~"democrat", partisan_identity =="republican"& concordance =="concordant"~"republican", partisan_identity =="democrat"& concordance =="discordant"~"republican", partisan_identity =="republican"& concordance =="discordant"~"democrat", TRUE~NA_character_ ),experiment_id =4,age = Age, # again, no direct info on year, we impute year of data collection reported for Study 3year =2020) ```### Combine and add identifiers (`country`, `paper_id`) and `news_selection`In study 2 and 3, the same news pool has been used. In study 4, a different pool has been used. We take this into account when merging the different studies. ```{r}## Combine + add remaining variablesbago_2022 <-bind_rows(d2, d3, d4) |>mutate(scale ="binary", country ="United States",paper_id ="bago_2022", # news id news_id =ifelse(experiment_id %in%c(2,3), news_id, paste0(experiment_id, "_", news_id) ), news_selection ="researchers", ) |># add_intervention_info bind_cols(intervention_info) |>select(any_of(target_variables))# check# bago_2022 |># group_by(news_id) |># summarize(n_observations = n()) |> # arrange(as.numeric(news_id))bago_2022 |>group_by(experiment_id, condition, intervention_label) |>count()```## Write out data```{r}save_data(bago_2022)```