Tag: Interdisciplinary studies

  • Gut microbiome strain-sharing within isolated village social networks

    Gut microbiome strain-sharing within isolated village social networks

    [ad_1]

    Local involvement and ethics

    We worked closely with the local population of Copan, sought approval and feedback from officials at the Ministry of Health (MOH) of Honduras, and endeavoured to provide practical benefits to the local community. When we began designing the underlying cohort project in 2013 (in 176 villages, including the 18 used here), the Bill and Melinda Gates Foundation introduced us to the Inter-American Development Bank (IDB), which has been supporting and doing work throughout Latin America, and the IDB in turn introduced us to the MOH. Because of this pathway to getting the project launched, we worked with local and regional public health agencies and with local leaders rather than with academic partners.

    The area we chose to work in the western highlands of Honduras, Copan, is very isolated. Over the years, as we built our data collection team in Copan, we developed deep ties to the local community, to local village leaders and to the few local health clinics there, as well as to local transportation and infrastructure providers. Because of these ties and our commitment to the local community, we presented our results directly to these constituencies regularly at the completion of our various projects.

    We provided other material benefits to the local community, beyond simply providing them with information. When we tested people for stool parasites, we gave them the results of their tests and arranged for them to be treated. When we tested people for vision, we provided corrective glasses. We solicited ideas from the local community about what infrastructure improvements we could make, and we repaired many local playgrounds and clinics as a result. We arranged for an American company to provide free portable handheld ultrasound devices to the local health clinics, which was much appreciated by local providers. In terms of capacity building, we hired and trained over 100 local people, and many of our former data collectors have gone on to work for other public health and development entities. Finally, we offered a talented young person from Copan a position as a PhD student in the USA.

    Throughout our work in Honduras, along with our extensive involvement at local and national levels, we have endeavoured to act with integrity, curiosity and respect in all our relationships.

    This research would not have been prohibited in the USA. This work is not likely to result in stigmatization, incrimination or discrimination or personal risk for the participants, and we have safeguarded all data from threats to the privacy or security of our participants.

    All participants provided informed consent, and our work was approved by the Yale Committee on Human Subjects (reference no. 2000020688).

    Network construction

    Village-level networks were mapped with standard ‘name generators’ for the whole village. After a photographic census (of all adolescent and adult residents) was taken for each village, we conducted the main network survey in each village, including a detailed, hour-long survey7, incorporating demographic and health measures, as well as a battery of name generators with which respondents identified relevant social relationships (friends, family members, people they spend free time with, and so on) through names and photographs shown in our TRELLIS software (available at trellis.yale.edu)45. All the name generator questions are listed in Supplementary Table 1.

    For questions in which a pair reported different levels of the same variable, such as greeting type or the amount of free time, we symmetrized the variables as follows: for greeting type, we reported the greeting type involving the most physical contact. For the frequency of free time and shared meals between a pair, we symmetrized by choosing the response that indicates more frequent contact. We symmetrized all other responses at the relationship level here (that is, when either of two people nominate each other as a ‘close friend’, we counted it). When calculating degree distributions, centralities and clustering, we simplified our networks to remove multiplexity (that is, we concatenated all ties between pairs of people) and symmetrized the ties (that is, we ignored who nominated whom in each pair).

    Social network graphs were analysed and geodesic distances and centrality measures were calculated with igraph (v.1.3.5)46 and plotted with the Fruchterman–Reingold algorithm. To protect the anonymity of our study villages, the villages were renamed to random town names from another country.

    Sample collection and sequencing

    Participants were instructed on how to self-collect the faecal samples using a training module delivered in person in the villages and were asked to return samples promptly to the local team. Samples were refrigerated immediately upon collection and then stored in liquid nitrogen at the collection site within 12 h after collection and moved to a −80 °C freezer in Copan Ruinas, Honduras. All the villages followed the same procedures. Samples were shipped, in randomized allotments, on dry ice to the USA and stored in −80 °C freezers.

    Stool material was homogenized using TissueLyzer from Qiagen, and the lysate was prepared for extraction with the Chemagic Stool gDNA extraction kit (Perkin Elmer) and extracted on the Chemagic 360 Instrument (Perkin Elmer) following the manufacturer’s protocol. Sequencing libraries were prepared using the KAPA Hyper Library Preparation kit (KAPA Biosystems). Shotgun metagenomic sequencing was carried out on Illumina NovaSeq 6000. Samples not reaching the desired sequencing depth of 50 Gbp were resequenced on a separate run. Raw metagenomic reads were deduplicated using prinseq lite47 (v.0.20.2) with default parameters. The resulting reads were screened for human contamination (hg19) with BMTagger and then quality filtered with Trimmomatic48 (v.0.36, parameters ‘ILLUMINACLIP: nextera_truseq_adapters.fasta:2:30:10:8:true SLIDINGWINDOW: 4:15 LEADING: 3 TRAILING: 3 MINLEN: 50’). This resulted in a total of 1,787 samples (with an average size of 8.6 × 107 reads).

    Species-level and strain-level profiling

    Species-level profiling was performed using MetaPhlAn 426 using the Jan21 database and default parameters. Strain-level profiling was performed for a subset of species present in at least 50 samples using StrainPhlAn 426 with parameters ‘–marker_in_n_samples 1 –sample_with_n_markers 10 — phylophlan_mode accurate’. This resulted in a total of 841 species-level genome bins (SGB) and 339,137 profiled strains. The StrainPhlAn ‘strain_transmission.py’ script was used to assess transmission events using the produced trees, which yielded a total of 513,177 identified events. For a robust calculation, strain-sharing rates were calculated only for pairs sharing at least ten SGBs.

    Beta diversity indices were calculated using the vegdist function from the vegan R package (v.2.6-2)49.

    Separation of distances by village membership was tested by permutational multivariate analysis of variance (PERMANOVA) using the adonis function from the vegan R package with 999 permutations.

    Statistical analyses

    All statistical analyses were performed in R (v.4.1.3). Correction for multiple testing (Benjamini–Hochberg procedure, marked Padj) was applied when appropriate, and significance was defined at Padj< 0.05. All tests were two-sided except where otherwise specified. All egocentric regressions (that is, when we assess the relationship of network position and strain-sharing) involved linear mixed-effects models with this general formula specification:

    $$\begin{array}{c}{\rm{O}}{\rm{u}}{\rm{t}}{\rm{c}}{\rm{o}}{\rm{m}}{\rm{e}}\,{\rm{o}}{\rm{f}}\,{\rm{i}}{\rm{n}}{\rm{t}}{\rm{e}}{\rm{r}}{\rm{e}}{\rm{s}}{\rm{t}} \sim {\rm{p}}{\rm{r}}{\rm{e}}{\rm{d}}{\rm{i}}{\rm{c}}{\rm{t}}{\rm{o}}{\rm{r}}\,{\rm{o}}{\rm{f}}\,{\rm{i}}{\rm{n}}{\rm{t}}{\rm{e}}{\rm{r}}{\rm{e}}{\rm{s}}{\rm{t}}+{\rm{a}}{\rm{g}}{\rm{e}}+{\rm{s}}{\rm{e}}{\rm{x}}\\ \,+\,{\rm{B}}{\rm{M}}{\rm{I}}+{\rm{B}}{\rm{r}}{\rm{i}}{\rm{s}}{\rm{t}}{\rm{o}}{\rm{l}}\,{\rm{s}}{\rm{t}}{\rm{o}}{\rm{o}}{\rm{l}}\,{\rm{s}}{\rm{c}}{\rm{a}}{\rm{l}}{\rm{e}}+{\rm{h}}{\rm{o}}{\rm{u}}{\rm{s}}{\rm{e}}{\rm{h}}{\rm{o}}{\rm{l}}{\rm{d}}\,{\rm{w}}{\rm{e}}{\rm{a}}{\rm{l}}{\rm{t}}{\rm{h}}\,{\rm{i}}{\rm{n}}{\rm{d}}{\rm{e}}{\rm{x}}\\ \,+\,{\rm{d}}{\rm{i}}{\rm{e}}{\rm{t}}\,{\rm{d}}{\rm{i}}{\rm{v}}{\rm{e}}{\rm{r}}{\rm{s}}{\rm{i}}{\rm{t}}{\rm{y}}\,{\rm{s}}{\rm{c}}{\rm{o}}{\rm{r}}{\rm{e}}+{\rm{m}}{\rm{e}}{\rm{d}}{\rm{i}}{\rm{c}}{\rm{a}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}\,{\rm{u}}{\rm{s}}{\rm{a}}{\rm{g}}{\rm{e}}+{\rm{w}}{\rm{a}}{\rm{t}}{\rm{e}}{\rm{r}}\,{\rm{s}}{\rm{o}}{\rm{u}}{\rm{r}}{\rm{c}}{\rm{e}}\\ \,+\,{\rm{D}}{\rm{N}}{\rm{A}}\,{\rm{c}}{\rm{o}}{\rm{n}}{\rm{c}}{\rm{e}}{\rm{n}}{\rm{t}}{\rm{r}}{\rm{a}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}+{\rm{s}}{\rm{e}}{\rm{q}}{\rm{u}}{\rm{e}}{\rm{n}}{\rm{c}}{\rm{i}}{\rm{n}}{\rm{g}}\,{\rm{d}}{\rm{e}}{\rm{p}}{\rm{t}}{\rm{h}}+{\rm{e}}{\rm{x}}{\rm{t}}{\rm{r}}{\rm{a}}{\rm{c}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}\,{\rm{d}}{\rm{a}}{\rm{t}}{\rm{e}}\\ \,+\,{\rm{s}}{\rm{h}}{\rm{i}}{\rm{p}}{\rm{p}}{\rm{i}}{\rm{n}}{\rm{g}}\,{\rm{b}}{\rm{a}}{\rm{t}}{\rm{c}}{\rm{h}}+{\rm{s}}{\rm{e}}{\rm{q}}{\rm{u}}{\rm{e}}{\rm{n}}{\rm{c}}{\rm{i}}{\rm{n}}{\rm{g}}\,{\rm{b}}{\rm{a}}{\rm{t}}{\rm{c}}{\rm{h}}+{\rm{e}}{\rm{x}}{\rm{t}}{\rm{r}}{\rm{a}}{\rm{c}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}\,{\rm{b}}{\rm{a}}{\rm{t}}{\rm{c}}{\rm{h}}\\ \,+\,(1|{\rm{v}}{\rm{i}}{\rm{l}}{\rm{l}}{\rm{a}}{\rm{g}}{\rm{e}})+(1|{\rm{b}}{\rm{u}}{\rm{i}}{\rm{l}}{\rm{d}}{\rm{i}}{\rm{n}}{\rm{g}})\end{array}$$

    That is, we controlled for age, sex, wealth, Bristol stool scale and body mass index (BMI), as well as sample properties (for example, DNA concentration) and village fixed effects. We also included household water source, individual medication usage in the last month and diet diversity (the number of food categories consumed on a daily basis10). Medication types included: painkillers, antibiotics, anti-diarrhoeal, anti-parasitic, anti-fungal, anti-diabetics, antacids, laxatives and vitamins. Mixed-effects models were created with the lmertest package (v.3.1.3)50.

    Network predictions

    Mixed-effects logistic regression models were used for out-of-sample network predictions. Class-balanced data sets were constructed by down-sampling the number of unrelated pairs to equal the number of related pairs, and we trained our model using k-fold cross-validation with k = 3, and predictions from the three separate test sets were combined. ROC curves were constructed from the average of five sets of threefold cross-validation. ROC curves and confidence intervals were calculated with the pROC package (v.1.18.0)51 and logistic regression models were constructed with the lmertest package (v.3.1.3) with the binomial family link function and a random slope per village.

    The predictive model including all covariates was specified by the following formula:

    $$\begin{array}{c}{\rm{R}}{\rm{e}}{\rm{l}}{\rm{a}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}{\rm{s}}{\rm{h}}{\rm{i}}{\rm{p}} \sim {\rm{m}}{\rm{i}}{\rm{c}}{\rm{r}}{\rm{o}}{\rm{b}}{\rm{i}}{\rm{o}}{\rm{m}}{\rm{e}}\,{\rm{s}}{\rm{i}}{\rm{m}}{\rm{i}}{\rm{l}}{\rm{a}}{\rm{r}}{\rm{i}}{\rm{t}}{\rm{y}}+{\rm{s}}{\rm{e}}{\rm{x}}\\ \,+\,{\rm{i}}{\rm{n}}{\rm{d}}{\rm{i}}{\rm{g}}{\rm{e}}{\rm{n}}{\rm{o}}{\rm{u}}{\rm{s}}\,{\rm{s}}{\rm{t}}{\rm{a}}{\rm{t}}{\rm{u}}{\rm{s}}+{\rm{r}}{\rm{e}}{\rm{l}}{\rm{i}}{\rm{g}}{\rm{i}}{\rm{o}}{\rm{n}}+{\rm{a}}{\rm{g}}{\rm{e}}\,{\rm{d}}{\rm{i}}{\rm{f}}{\rm{f}}{\rm{e}}{\rm{r}}{\rm{e}}{\rm{n}}{\rm{c}}{\rm{e}}\\ \,+\,{\rm{a}}{\rm{v}}{\rm{e}}{\rm{r}}{\rm{a}}{\rm{g}}{\rm{e}}\,{\rm{a}}{\rm{g}}{\rm{e}}+{\rm{w}}{\rm{e}}{\rm{a}}{\rm{l}}{\rm{t}}{\rm{h}}\,{\rm{d}}{\rm{i}}{\rm{f}}{\rm{f}}{\rm{e}}{\rm{r}}{\rm{e}}{\rm{n}}{\rm{c}}{\rm{e}}+{\rm{a}}{\rm{v}}{\rm{e}}{\rm{r}}{\rm{a}}{\rm{g}}{\rm{e}}\,{\rm{w}}{\rm{e}}{\rm{a}}{\rm{l}}{\rm{t}}{\rm{h}}\\ \,+\,{\rm{e}}{\rm{d}}{\rm{u}}{\rm{c}}{\rm{a}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}\,{\rm{d}}{\rm{i}}{\rm{f}}{\rm{f}}{\rm{e}}{\rm{r}}{\rm{e}}{\rm{n}}{\rm{c}}{\rm{e}}+{\rm{a}}{\rm{v}}{\rm{e}}{\rm{r}}{\rm{a}}{\rm{g}}{\rm{e}}\,{\rm{e}}{\rm{d}}{\rm{u}}{\rm{c}}{\rm{a}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}\\ \,+\,{\rm{m}}{\rm{e}}{\rm{d}}{\rm{i}}{\rm{c}}{\rm{a}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}\,{\rm{u}}{\rm{s}}{\rm{a}}{\rm{g}}{\rm{e}}+{\rm{s}}{\rm{a}}{\rm{m}}{\rm{e}}\,{\rm{w}}{\rm{a}}{\rm{t}}{\rm{e}}{\rm{r}}\,{\rm{s}}{\rm{o}}{\rm{u}}{\rm{r}}{\rm{c}}{\rm{e}}+{\rm{d}}{\rm{i}}{\rm{e}}{\rm{t}}\\ \,+\,{\rm{B}}{\rm{r}}{\rm{i}}{\rm{s}}{\rm{t}}{\rm{o}}{\rm{l}}\,{\rm{s}}{\rm{t}}{\rm{o}}{\rm{o}}{\rm{l}}\,{\rm{s}}{\rm{c}}{\rm{a}}{\rm{l}}{\rm{e}}+{\rm{h}}{\rm{o}}{\rm{u}}{\rm{s}}{\rm{e}}{\rm{h}}{\rm{o}}{\rm{l}}{\rm{d}}\,{\rm{s}}{\rm{h}}{\rm{a}}{\rm{r}}{\rm{i}}{\rm{n}}{\rm{g}}\\ \,+\,(0+{\rm{m}}{\rm{i}}{\rm{c}}{\rm{r}}{\rm{o}}{\rm{b}}{\rm{i}}{\rm{o}}{\rm{m}}{\rm{e}}\,{\rm{s}}{\rm{i}}{\rm{m}}{\rm{i}}{\rm{l}}{\rm{a}}{\rm{r}}{\rm{i}}{\rm{t}}{\rm{y}}|{\rm{v}}{\rm{i}}{\rm{l}}{\rm{l}}{\rm{a}}{\rm{g}}{\rm{e}}\,{\rm{I}}{\rm{D}})\end{array}$$

    where ‘microbiome similarity’ is either the strain-sharing rate, Jaccard index or Bray–Curtis dissimilarity calculated between the members of a pair.

    Variable importance metrics were calculated based on the permutation feature importance metric using the car R package (v.3.0). The permutation feature importance is defined to be the decrease in a model score when a single feature value is shuffled randomly52. This procedure breaks the relationship between the feature and the target; thus, the drop in the model score is indicative of how much the model depends on the feature. Variable importance metrics were analysed after 1,000 random permutations of each feature. Variable inflation factor values were calculated to ensure the reliability of results against collinearity of variables and were all low (less than 2).

    Microbiome null permutations

    Microbiome null permutations create a null distribution of strain-sharing rates between any two people while accounting for (just) the network structure. Under the null hypothesis that a host’s microbiome composition and social network are independent, we can sever their relationship by randomly permuting the microbiome of every person in the village and recalculating metrics of interest, for example, strain-sharing by degree or clustering Rand indices. This ensures that the inherent structural pattern of the network remains the same, but the node values are randomized. This allows us to observe the distribution of our statistics if the human microbiome is fostered independently of any host social interactions.

    Village-wide microbiome permutations were used to calculate null distributions for the strain-sharing rate by geodesic distance and for the clustering results. For relationship-specific permutations in Supplementary Fig. 1, permutations at the relationship level were taken instead of full village permutations. The observed distribution of relationship-specific sharing was compared with the distribution of sharing observed when that specific relationship tie was permuted, for example, comparing the sharing between someone and their friend versus someone and 100 random people’s friends in the same village. For the inherently gendered relationships of husband/wife and mother/father of a child, we accounted for the sex of the ego, but for all other relationships that are not necessarily gendered (for example, free time), we did not.

    Longitudinal analyses

    A subset of 301 people from four villages were followed-up after a period of 2 years and asked to provide a second stool sample. Samples were processed consistently with the same pipeline used to analyse the previously processed 1,787 samples.

    We defined relationship ties by using the same social network from the initial wave and evaluated the following linear mixed-effect model formula:

    $${\rm{SS}}{{\rm{R}}}_{{\rm{T}}2} \sim SS{R}_{{\rm{T}}1}+{\rm{relationship}}+M+(1| {\rm{village}}\,{\rm{ID}})+(1| {\rm{ego}})$$

    where SSRT1 and SSRT2 are the strain-sharing rate in pairs of people at time points T1 and T2, respectively. We show standardized coefficients.

    To decompose the effect of sharing across all species, we used a mixed-effect logistic model formula specified as follows:

    $${\rm{T}}{2}_{S} \sim {\rm{T}}{1}_{S}+{\rm{relationship}}+M+(1| {\rm{species}})+(1| {\rm{villageID}})+(1| {\rm{ego}})$$

    where \({\rm{T}}{1}_{S}\) and \({\rm{T}}{2}_{S}\) are binary variables indicating whether we observed strain-sharing of an individual species at time point T1 or T2, for all species combined. A random intercept for each individual species was added as well as for village membership and person.

    In both models, ‘relationship’ is a dummy variable indicating the presence (or absence) of a tie between the pair of people, and M is the Mahalanobis distance calculated on the following covariates:

    $$\begin{array}{c}M={\rm{M}}{\rm{a}}{\rm{h}}{\rm{a}}{\rm{l}}{\rm{a}}{\rm{n}}{\rm{o}}{\rm{b}}{\rm{i}}{\rm{s}}({\rm{a}}{\rm{g}}{\rm{e}},{\rm{s}}{\rm{e}}{\rm{x}},{\rm{B}}{\rm{M}}{\rm{I}},{\rm{B}}{\rm{r}}{\rm{i}}{\rm{s}}{\rm{t}}{\rm{o}}{\rm{l}}\,{\rm{s}}{\rm{t}}{\rm{o}}{\rm{o}}{\rm{l}}\,{\rm{s}}{\rm{c}}{\rm{a}}{\rm{l}}{\rm{e}},\\ \,\,{\rm{h}}{\rm{o}}{\rm{u}}{\rm{s}}{\rm{e}}{\rm{h}}{\rm{o}}{\rm{l}}{\rm{d}}\,{\rm{w}}{\rm{e}}{\rm{a}}{\rm{l}}{\rm{t}}{\rm{h}}\,{\rm{i}}{\rm{n}}{\rm{d}}{\rm{e}}{\rm{x}},{\rm{d}}{\rm{i}}{\rm{e}}{\rm{t}}\,{\rm{d}}{\rm{i}}{\rm{v}}{\rm{e}}{\rm{r}}{\rm{s}}{\rm{i}}{\rm{t}}{\rm{y}}\,{\rm{i}}{\rm{n}}{\rm{d}}{\rm{e}}{\rm{x}},\\ \,\,{\rm{m}}{\rm{e}}{\rm{d}}{\rm{i}}{\rm{c}}{\rm{a}}{\rm{t}}{\rm{i}}{\rm{o}}{\rm{n}}\,{\rm{u}}{\rm{s}}{\rm{a}}{\rm{g}}{\rm{e}},{\rm{w}}{\rm{a}}{\rm{t}}{\rm{e}}{\rm{r}}\,{\rm{s}}{\rm{o}}{\rm{u}}{\rm{r}}{\rm{c}}{\rm{e}},{\rm{b}}{\rm{u}}{\rm{i}}{\rm{l}}{\rm{d}}{\rm{i}}{\rm{n}}{\rm{g}}\,{\rm{I}}{\rm{D}})\end{array}$$

    The pairwise Mahalanobis distance was calculated on the covariates matrix using the D2.dist function from the biotools R package53 (v.4.2). We also specified this model using the constituent variables, rather than the Mahalonobis distance (Supplementary Data 2).

    Microbiome and social clustering

    We use the Louvain and the Leiden methods as implemented in the igraph package to cluster participants along social and microbiome lines. Louvain clustering is based on greedy modularity optimization. Modularity is a scale value between −0.5 (non-modular clustering) and 1 (fully modular clustering) that measures the relative density of edges inside communities compared with edges outside communities. Optimizing this value theoretically results in the best possible grouping of the nodes of a given network. In cases where a pair shared too few SGBs to calculate a robust strain-sharing rate (fewer than ten), a strain-sharing rate of 0% was imputed to allow for proper weight-based clustering. This occurred in 0.45% of the pairwise comparisons (16,228 out of 3,560,769 comparisons), and just 838 of the 16,228 comparisons were from people in the same village. The adjusted Rand index was calculated with the mclust package (v.6.0.0)54.

    For testing species differential abundance across network communities with the Kruskal–Wallis test, robustness checks ensuring that each social cluster had more than five people and the species was present in more than five people in the village were performed, and cases where this criterion was not met were excluded.

    Reporting summary

    Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.

    [ad_2]

    Source link

  • Mortality caused by tropical cyclones in the United States

    [ad_1]

    Data

    Wind speed data

    We measure TC incidence using the LICRICE model (version 4)28. LICRICE is a parametric wind-field model that estimates the maximum sustained winds experienced by every location throughout the lifetime of each TC recorded in the International Best Track Archive for Climate Stewardship (IBTrACS) database2,29,31,37,53,54,55,56. LICRICE uses observed maximum wind speeds to reconstruct wind fields throughout the storm based on internal storm structure, location, and each storm’s observed translational velocity. Many summaries of storm wind incidence are possible to construct using LICRICE, such as integrated power dissipation28,29; however, prior analyses have shown that the maximum wind experienced within storms is the most predictive of social and economic outcomes among numerous parsimonious metrics previously analysed. This result is consistent with most components of built structures failing catastrophically based on whether or not a threshold of stress is applied. LICRICE does not explicitly model storm surge, rainfall or flooding, however these dimensions of impact are captured in the analysis to the extent that they are correlated with integrated maximum wind speeds. For example, we find that LICRICE wind speed and NCEP rainfall from TCs are correlated within storms at the pixel level (Supplementary Fig. 3) and at the state-by-storm level (R2 = 0.31, P < 0.001; Fig. 1e) for a limited sample of storms for which granular rainfall data are available. Iterations of the LICRICE model has been used to measure various social and economic impacts of TCs, including direct deaths and damages29, changes to household income and expenditure2, infant mortality2, GDP growth28,37 and depreciation3. We also compare our measure of wind speed against total direct national economic damages (normalized by GDP) from a limited sample of TCs estimated by Nordhaus51. We find that national wind speed exposure is a meaningful predictor of total national damages (Fig. 1d and Supplementary Fig. 2), although we note that this outcome is highly uncertain and widely understood to be biased.

    Here we use a new reconstruction of incidence at the sub-national-level within CONUS. We reconstruct incidence for 0.125° × 0.125° pixel of CONUS in each of 1,230 Atlantic storms between 1930 and 2015. Supplementary Fig. 1 shows all decadal averages of these output (four example maps are also shown in Fig. 1a), illustrating the TC climatology for CONUS–however these aggregates over time are not themselves used in subsequent analysis.

    To match state-level mortality data, TC incidence is collapsed from pixels to states for every month. If multiple storms impact a cell within a month, the maximum incidence at the cell level is recorded, and monthly averages are computed across pixels in each state. This spatial averaging causes our measures of incidence to be substantially lower than the maximum sustained wind speed commonly reported for storms, since only a small number of pixels experience those extreme conditions within each storm. We note that TC events generate the highest average state wind speeds compared to wind speeds from other intense storm phenomena, such as tornadoes. For reference, the minimum monthly state average wind speed we compute from TCs in our sample is 3.34 × 10−4 ms1 and the 1st percentile is 8.4 × 10−3 ms−1. By contrast, the maximum monthly state average wind speeds from tornadoes in CONUS between 1950 and 2022 is 9.6 × 10−4 ms1 and the 99th percentile is 3.8 × 10−5 ms1, and for non-TC and non-tornado wind/hail events the maximum is 1.1 × 10−3 ms1 and the 99th percentile is 1.3 × 10−4 ms1. Therefore, the maximum non-TC wind events are comparable to the minimum (non-zero) TC events; and absent a TC, states do not experience average wind speeds of a similar magnitude as those from TCs.

    Prior analysis by Hsiang & Jina3 demonstrated that spatial aggregates of TC exposure can be used as independent variables in a regression framework to obtain unbiased average effects that are expressed at finer spatial resolutions (footnote 13 on pages 16–17 of ref. 3). As long as there is no systematic correlation between the average intensity of a storm and the likelihood that the most intense regions within that storm strike the most populated (or economically active or vulnerable) pixels within a state, regression coefficients will not be biased by spatial aggregations. This condition would be violated if, for example, there were systematic patterns such that the eyes of a Category 3 hurricanes tended to pass directly over dense cities, but the eyes of Category 2 hurricanes tended to miss cities. However, given that the paths of storms are primarily controlled by random steering winds at high altitude, interacting with the beta-effect induced by the Earth’s meridional vorticity gradient, we have strong reason to believe that the spatial distribution of TC incidence within each state is orthogonal to the spatial distribution of underlying populations; and further that this covariance is independent of average TC intensity. Thus far, we know of no evidence that the trajectory of stronger (or weaker) storms systematically strike more vulnerable locations on land.

    Of the 1,230 TCs that we reconstruct, 501 come within 250 km of a CONUS coastline. Intersecting these storms with state boundaries generates a total of 3,317 state-by-TC events. These longitudinal data reveals rich variation in the timing and intensity of TC incidence for individual states52 (Extended Data Fig. 1). Within-state variation in incidence season-to-season and month-to-month provides substantial variation in TC impulses that enable us to identify the impulse-response of mortality empirically.

    All-cause mortality data

    We analyse all-cause mortality at the state-year-month level between 1930 and 2015 using data from multiple sources. Data from 1900 to 2004 were digitized and assembled by Barreca et al.31 in their report identifying the impact of temperature on mortality. According to Barreca, this is the most comprehensive data on mortality assembled in this context. The remaining data was assembled by the authors using the CDC Underlying Cause of Death database. Data prior to 1959 was digitized from the Mortality Statistics of the United States annual volumes and is not otherwise available in a machine-readable format. Therefore, data in years prior to 1959 do not include cause of death (for example, cardiovascular disease) or demographic information (for example, age 1–44, Black)31. From 1959 to 2004, the data are from the machine-readable MCOD files, which include cause of death and demographic data.

    For the years 2005–2015, we analyse mortality data from the public CDC Underlying Cause of Death database (2017). The data are based on death certificates for U.S. residents, which gives a single underlying cause of death and demographic data. Cause of death prior to 2000 was indexed using the four-digit ICD-9 code and 2000 onwards the index changed to the four-digit ICD-10 code. Cause of death was indexed using a four-digit ICD-10 code. We harmonized the cause of death into five categories that matched the cause of death variables from Barreca et al. We also construct a 6th category which is the difference between all-cause mortality and the sum of the 5 cause-specific categories, called ‘other’. Notably, the change in CDC ICD code methodology resulted in a shift in the counts of deaths from specific causes, particularly infectious diseases and cardiovascular disease.

    To account for differences in underlying age-specific mortality we decomposed the effect of TCs on all-cause mortality by four age groups in the data: <1, 1–44, 45–64, and 65+ years of age. We were limited to these age groups because these are the designations in our historical data. We computed mortality with respect to the underlying population by these same four age groups. We compute mortality by race with respect to the population by race in each state and year. Black and white are the only race categories available for the entire sample. Extended Data Fig. 3 shows the monthly all-cause mortality rate, and our predicted monthly mortality rate, for all the states in our sample.

    Direct deaths from TCs

    Direct deaths from TCs are deaths that are officially attributed to a storm by the US government. We combine official death counts from two NOAA data sources. For storms between 1950 and 1996 we use the NOAA National Hurricane Center and Central Pacific Hurricane Center’s Hurricane in History6. Storms from 1997 to 2015 are from the NOAA Storm Events Database7.

    Population data

    We normalize state mortality by the population (per 100,000 people) in the state each month. Similar to the all-cause mortality data, these data must be combined from multiple sources. Pre-1968 population estimates are from Haines57; estimates for 1969–2000 are from the National Cancer Institute (2008); estimates for 2000–2010 are from the US Census Bureau, Intercensal Population and Housing Unit Estimates: 2000 to 2010 (ref. 58); estimates for 2010–2017 are from the US Census Bureau, US Population Estimates59.

    Temperature data

    Average monthly temperature data are from Berkeley Earth Surface Temperatures (BEST) land surface air temperature. BEST provides a monthly mean of average, minimum and maximum surface air temperature over land covering 1753 to the present56,60. The temperature data are based on a large inventory of observations from over 30,000 weather stations. Using these observations gridded temperature fields are reconstructed statistically, incorporating the reliability of individual weather stations and spatial variability of temperature56,60. Gridded BEST temperature data are then spatially aggregated, weighted by population, to the state-month-level.

    Analysis

    The econometric approach that we apply here is a top-down strategy, commonly called a ‘reduced-form’ analysis, that describes the overall net change of an aggregate outcome y (mortality) in response to exogenous treatments z (TC incidence). Under suitable conditions, this approach can identify causal effects on the outcome y induced by exogenous changes in independent variable z without explicitly describing all underlying mechanisms that link z to y, without observing intermediary variables x (for example, retirement savings accounts or healthcare infrastructure) that might link z to y, or without explicitly tracking other determinants of y unrelated to z (such as demographic trends or health policy), denoted w (refs. 23,61,62). Let f(•) describe a complex and unobserved process that generates state-level mortality rates \({y}_{{t}_{2}}\), occurring at time t2 based on x, w and z that occur both at times t1 and t2 (t1 < t2):

    $${y}_{{t}_{2}}=f({z}_{{t}_{2}},{x}_{{t}_{1}}^{1}({z}_{{t}_{1}}),\ldots ,{x}_{{t}_{1}}^{K}({z}_{{t}_{1}}),\,{x}_{{t}_{2}}^{1}({z}_{{t}_{1}},{z}_{{t}_{2}}),\ldots ,{x}_{{t}_{2}}^{K}({z}_{{t}_{1}},{z}_{{t}_{2}}),\,{w}_{{t}_{1}\,}^{1}\ldots \,{w}_{{t}_{1}\,}^{J},\,{w}_{{t}_{2}\,}^{1}\ldots \,{w}_{{t}_{2}}^{J})$$

    (1)

    where \({x}_{{t}_{1}}^{k}\left({z}_{{t}_{1}}\right)\) indicates that the kth factor xk, which influences mortality rates y, at time t1 is itself affected by TCs at time t1. At time t2, xk may be influenced both by TCs at t2 and those that occur in the past at t1. Here, we let there be K pathways through which y is impacted by intermediary variables (x) and J ways through which determinants unrelated to TCs (w) impact y.

    In this framework, the direct mortality impact of TC incidence usually reported by government agencies are the partial derivative:

    $$direct{\rm{\_}}{deaths}_{{t}_{2}}=\frac{{\rm{\partial }}{y}_{{t}_{2}}}{{\rm{\partial }}{z}_{{t}_{2}}}$$

    (2)

    which are the deaths that occur contemporaneously and directly as a result of the geophysical event itself, holding fixed all other factors.

    In this analysis, we directly estimate the total change in mortality that results from TC incidence in both current and prior moments in time, allowing for the possibility that changes in other factors that are influenced by TCs may indirectly affect mortality. In this case, the overall total change in mortality from TC incidence in t2 is the total derivative:

    $$\frac{{\rm{d}}}{{\rm{d}}{z}_{{t}_{2}}}({mortality{\rm{\_}}rate}_{{t}_{2}})=\frac{{\rm{d}}{y}_{{t}_{2}}}{{\rm{d}}{z}_{{t}_{2}}}=\mathop{\underbrace{\frac{{\rm{\partial }}{y}_{{t}_{2}}}{{\rm{\partial }}{z}_{{t}_{2}}}}}\limits_{{\rm{d}}{\rm{i}}{\rm{r}}{\rm{e}}{\rm{c}}{\rm{t}}\,{\rm{d}}{\rm{e}}{\rm{a}}{\rm{t}}{\rm{h}}{\rm{s}}}+\mathop{\underbrace{\mathop{\sum }\limits_{k=1}^{K}\frac{{\rm{\partial }}{y}_{{t}_{2}}}{{\rm{\partial }}{x}_{{t}_{2}}^{k}}\frac{{\rm{\partial }}{x}_{{t}_{2}}^{k}}{{\rm{\partial }}{z}_{{t}_{2}}}}}\limits_{{\rm{i}}{\rm{n}}{\rm{d}}{\rm{i}}{\rm{r}}{\rm{e}}{\rm{c}}{\rm{t}}\,{\rm{d}}{\rm{e}}{\rm{a}}{\rm{t}}{\rm{h}}{\rm{s}}}$$

    (3)

    which includes both direct deaths and deaths that result from any of the K possible pathways that depend on the intermediate variables xk at time t2. Empirically, we find that direct deaths are much smaller than indirect deaths in CONUS.

    In addition, we also account for the possibility that deaths are delayed relative to TC incidence. Because direct deaths are usually tabulated immediately following storms, there are negligible direct deaths that are delayed. However, once we begin considering indirect deaths, it becomes possible for substantial delays to emerge due to the dynamics of different pathways xk. In our analysis, we also estimate the total deaths that occur at time t2 as a result of TC incidence that occurs at an earlier time t1, which is the total derivative

    $$\frac{{\rm{d}}}{{\rm{d}}{z}_{{t}_{1}}}({mortality{\rm{\_}}rate}_{{t}_{2}})=\frac{{\rm{d}}{y}_{{t}_{2}}}{{\rm{d}}{z}_{{t}_{1}}}=\mathop{\underbrace{\mathop{\sum }\limits_{k=1}^{K}\frac{{\rm{\partial }}{y}_{{t}_{2}}}{{\rm{\partial }}{x}_{{t}_{2}}^{k}}\frac{{\rm{\partial }}{x}_{{t}_{2}}^{k}}{{\rm{\partial }}{z}_{{t}_{1}}}}}\limits_{{\rm{i}}{\rm{n}}{\rm{d}}{\rm{i}}{\rm{r}}{\rm{e}}{\rm{c}}{\rm{t}}\,{\rm{v}}{\rm{i}}{\rm{a}}\,{x}_{{t}_{2}}}+\mathop{\underbrace{\mathop{\sum }\limits_{k=1}^{K}\frac{{\rm{\partial }}{y}_{{t}_{2}}}{{\rm{\partial }}{x}_{{t}_{1}}^{k}}\frac{{\rm{\partial }}{x}_{{t}_{1}}^{k}}{{\rm{\partial }}{z}_{{t}_{1}}}}}\limits_{{\rm{i}}{\rm{n}}{\rm{d}}{\rm{i}}{\rm{r}}{\rm{e}}{\rm{c}}{\rm{t}}\,{\rm{v}}{\rm{i}}{\rm{a}}\,{x}_{{t}_{1}}}$$

    (4)

    This expression does not contain a term for direct deaths, but it contains two summations which capture the effects of past TC incidence \(({{z}_{t}}_{1})\) on current mortality \((\,{y}_{{t}_{2}})\) via past intermediate variables \(({x}_{{t}_{1}})\) and current intermediate variables \(({x}_{{t}_{2}})\). In practice, we explore the possibility of indirect effects that emerge over the course of 240 months following TC incidence, one could generalize this framing to a corresponding number of summations.

    The possibility of delayed indirect deaths has two major implications regarding how indirect mortality is estimated and how those results are interpreted. First, because TC incidence at multiple points in the past, as well as the present, might affect current mortality, we must account for both the present and past influence of TC incidence simultaneously for each instance of the outcome. This is accomplished via deconvolution25,26,27,56,63,64, implemented here using a distributed lag-model solved via ordinary least squares, detailed below.

    Second, each TC event affects mortality outcomes at multiple points in time, thus computing the full impact of a TC event requires summing these impacts that might emerge gradually. In the simplified two-period framework above, the total impact from TC incidence at t1 is then

    $$\frac{{\rm{d}}}{{\rm{d}}{z}_{{t}_{1}}}(mortality{\rm{\_}}{rate}_{{t}_{1}}\,+\,mortality{\rm{\_}}{rate}_{{t}_{2}})\,=\,\frac{{\rm{d}}{y}_{{t}_{1}}}{{\rm{d}}{z}_{{t}_{1}}}\,+\,\frac{{\rm{d}}{y}_{{t}_{2}}}{{\rm{d}}{z}_{{t}_{1}}}$$

    (5)

    which can be expanded further by substituting from the equations above. These terms, if plotted separately, characterize the impulse response of y in reaction to the TC impulse \({z}_{{t}_{1}}\). In our actual analysis, we compute the average cumulative impact of a single TC that occurs at t0 over 240 subsequent months (t1 − t240). Following substitution and simplification, this can be expressed as

    $$\frac{{\rm{d}}}{{\rm{d}}{z}_{{t}_{0}}}\left(\mathop{\sum }\limits_{\ell =0}^{240}{mortality{\rm{\_}}rate}_{{t}_{\ell }}\right)=\mathop{\underbrace{\frac{{\rm{\partial }}{y}_{{t}_{0}}}{{\rm{\partial }}{z}_{{t}_{0}}}}}\limits_{{\rm{d}}{\rm{i}}{\rm{r}}{\rm{e}}{\rm{c}}{\rm{t}}\,{\rm{d}}{\rm{e}}{\rm{a}}{\rm{t}}{\rm{h}}{\rm{s}}}+\,\mathop{\underbrace{\mathop{\sum }\limits_{\ell =0}^{240}\mathop{\sum }\limits_{k=1}^{K}\left(\mathop{\sum }\limits_{j=\ell }^{240}\frac{{\rm{\partial }}{y}_{{t}_{j}}}{{\rm{\partial }}{x}_{{t}_{\ell }}^{k}}\right)\frac{{\rm{\partial }}{x}_{{t}_{\ell }}^{k}}{{\rm{\partial }}{z}_{{t}_{0}}}}}\limits_{{\rm{i}}{\rm{n}}{\rm{d}}{\rm{i}}{\rm{r}}{\rm{e}}{\rm{c}}{\rm{t}}\,{\rm{d}}{\rm{e}}{\rm{a}}{\rm{t}}{\rm{h}}{\rm{s}}}$$

    (6)

    which describes the overall total impact of a storm through all pathways across all possible delays  = [0, 240]. Note that neither K nor xk need ever be specified explicitly in our estimation below. This expansion reveals that, accounting for numerous possible pathways operating over different delays, a single TC event can potentially generate a total mortality impact much larger than the direct deaths traditionally reported.

    To compute the overall mortality burden imposed by the TC climate of CONUS, we compute the full mortality response across all age groups in each state, accounting for the incidence of each storm on the state:

    $$\begin{array}{l}mortality{\rm{\_}}burden\,=\,\mathop{\sum }\limits_{{\ell }=0}^{240}\sum _{t\in {\rm{m}}{\rm{o}}{\rm{n}}{\rm{t}}{\rm{h}}{\rm{s}}}\sum _{s\in {\rm{s}}{\rm{t}}{\rm{o}}{\rm{r}}{\rm{m}}{\rm{s}}}\sum _{i\in {\rm{s}}{\rm{t}}{\rm{a}}{\rm{t}}{\rm{e}}{\rm{s}}}{population}_{i,t+{\ell }}\\ \,\,\,\,\,\cdot {z}_{sit}\cdot \frac{{\rm{d}}}{{\rm{d}}z}{(mortality{\rm{\_}}rate)}_{i,t+{\ell }}\end{array}$$

    (7)

    where zsit is the TC incidence of storm s on state i in month t, populationi,t+ is the population in state i in month t + , and \(\frac{{\rm{d}}}{{\rm{d}}z}{(mortality{\rm{\_}}rate)}_{i,t+{\ell }}\) is our estimate for the total impact of TC incidence on the mortality rate in state i in month t + . In practice, this effect is nonlinear, but it is expressed linearly here for simplicity. Information on state i affects the impulse responses used in these calculations because TC risk is computed by state and affects the structure of the impulse-response function.

    Econometric implementation

    Identification

    Our econometric analysis exploits the quasi-random variation in the location and intensity of TC incidence to estimate the impact of TCs on mortality separately from other known and unknown factors that affect mortality across locations and over time. As described above, this reduced-form approach captures the effect of all possible channels of influence that may increase mortality after a TC33,65. Because the location, timing and intensity of TC incidence is determined by oceanic and atmospheric conditions that are beyond the control of individual states, we assume mortality TC incidence is as good as randomly assigned23,61. For reference, Extended Data Fig. 1 shows the sequence of monthly TC incidence by state for all the states in our sample.

    We note that some early analyses of natural disaster impacts utilized social outcomes (for example, direct economic damage66 or direct mortality) as a proxy measure of physical hazard severity. However, it is now understood that use of these metrics as independent variables may confound estimated treatment effects, since they are endogenously determined by many of the same underlying covariates (for example, healthcare, infrastructure, inequality and institutions) that mediate other outcomes from disasters22,29,33,34,35,67. Thus, use of these proxy measures for hazard severity exposes analyses to selection biases, since population characteristics may cause observational units to ‘select’ into more or less severe treatment23. We therefore focus this analysis strictly on independent variables that are physical measures of TC incidence (wind speed), because they are exogenous and cannot be influenced by the populations that are impacted22.

    Deconvolution

    In considering the long-run impact of TCs on mortality, we hypothesize that there may be a delay between the geophysical event and components of the mortality response. Because TCs are regular events that occur frequently in CONUS, the possibility of this delay means that the time series of mortality outcomes we observe in data may be the result of overlapping responses from multiple storms. Extended Data Fig. 2 displays a cartoon of this data-generating process. In such a context, the empirical challenge is isolating the impact from individual storms which might be partially confounded by the overlapping TC signals from earlier or later storms. We use the well-established signal-processing approach of deconvolution25,26,27,33,63,64 to recover the characteristic impulse-response function for a TC impulse. Conceptually, this approach searches for an impulse-response function that, if applied to all TCs in the data simultaneously, best fits the observed outcome data. Stated another way, this approach estimates the effect of each TC accounting for the potential overlapping impact of all other TCs, subject to the constraint that TCs share a characteristic impulse-response function.

    This method assumes that the overlapping responses influencing mortality at a moment in time are additively separable, an assumption that we think is reasonable given the overall small impact that any individual storm event has mortality rates at a moment in time in a particular region (0.019% on average, 0.04% for storms at the 95th percentile). We solve for the structure of the impulse response, characterized by a set of coefficients β, using ordinary least squares (OLS). This is a standard procedure that is commonly applied in a wide range of disciplines26. In some fields, such as econometrics, deconvolution is frequently described as estimation of distributed lags27.

    Baseline specification

    Our main results are based on a linear model of TC incidence on mortality rate. Indexing states by i and month of sample by t, we solve the model

    $$\begin{array}{l}mortality{\rm{\_}}{rate}_{it}=\mathop{\sum }\limits_{{\ell }=-72}^{240}({\beta }_{{\ell }}\cdot wind{\rm{\_}}{speed}_{i,t-{\ell }})\\ \,\,\,\,\,\,\,\,+{\delta }_{1,i}\cdot {temp}_{it}\cdot {s}_{i}+\,{\delta }_{2,{\rm{i}}}\cdot {temp}_{it}^{2}\cdot {s}_{i}+{\mu }_{1}\cdot {m}_{it}\\ \,\,\,\,\,\,\,\,+\,\mathop{\sum }\limits_{n=1}^{8}({\eta }_{n}\cdot {s}_{i}\cdot {t}^{n})\,+\,{\mu }_{2}\cdot {m}_{it}\cdot t+{\mu }_{3}\cdot {h}_{t}+{{\epsilon }}_{it}\end{array}$$

    (8)

    via OLS. Here wind_speedi,t−ℓ is TC maximum wind speed months prior to month t, si is a vector of state-specific dummies, mit are state-by-month-of-the-year dummies (for example, an indicator variable for whether state = Florida and month = January), ht are month-of-sample dummies (for example, an indicator variable for whether the month = January, 1974), \({temp}_{it}\cdot {s}_{i}\) is month-of-sample temperature interacted with state dummies, and \({temp}_{it}^{2}\cdot {s}_{i}\) is squared month-of-sample temperature (also interacted with state). Each coefficient β measures the marginal effect of an additional ms1 of wind speed incidence on mortality months after a TC conditional on the effect of any prior TC. We include 72 lead terms in equation (8) as a falsification test, also known as negative exposure controls, since idiosyncratic future TC incidence should not alter current health outcomes.

    This model accounts for state-specific quadratic effects of temperature on mortality based on prior literature, which has shown that very hot and very cold temperatures cause higher levels of mortality relative to more moderate temperatures31,33,34,35. Each state is allowed to express a different mortality response to temperature extremes, implemented via interaction with the state dummy variable si. Supplementary Fig. 6 shows the state-specific shape of the quadratic functions we estimate for the temperature-mortality response. Consistent with prior findings studying patterns of adaptation31,33,34,35, we observe that some states have a flatter response at temperatures that are more common for that state (for example, cold in Minnesota) while other states have steeper curves at those same temperatures if they are less common (for example, cold in Florida). In an effort to balance parsimony with model richness, we omit extended lags of temperature based on prior literature demonstrating that impacts on mortality dissipate within a month33,68.

    This model also non-parametrically accounts for:

    • State-by-month-specific constants (fixed effects) that capture average differences between states, as well as unique seasonable patterns within states (\({\mu }_{1}\cdot {m}_{it}\)). These terms will account for differences in mortality driven by unobserved factors at the state level, such as health policies, as well as factors that cause seasons within a state to exhibit higher mortality, such as holidays.

    • State-specific nonlinear trends in mortality, captured by eighth-order polynomials in month-of-sample interacted with state fixed effects (\({\sum }_{n=1}^{8}({\eta }_{n}\cdot {s}_{i}\cdot {t}^{n})\)). These trends account for unobserved factors that have caused mortality within states to change over time, such as changing health policies or demographic trends.

    • Trends in state-specific seasonal patterns of mortality, captured by a linear trend in month-of-sample interacted with state and month fixed effects (\({\mu }_{2}\cdot {m}_{t}\cdot t\)). These trends are additive to the state-specific polynomial and allow for the model to express gradual convergence or divergence in the seasonality of mortality within a year, and allows for these changes to differ by state. These trends account for unobserved factors that drive gradual changes over time that may cause mortality in certain times of year (for example, January) to change relative to other times of year (for example, June). For example, if adoption of safety standards has reduced wintertime mortality from motor vehicle accidents or improvements in medical care have reduces summertime deaths from infectious diseases. Extended Data Fig. 4a illustrates the combined effect of these state-specific seasonal trends, state-specific polynomials, and state-by-month- specific constants on model predictions for Florida and New Jersey. For example, the seasonality of mortality in Florida has lessened over time, in conjunction with other nonlinear trends.

    • National month-of-sample fixed effects that capture nonlinear and/or discontinuous changes in mortality rates nation-wide (\({\mu }_{3}\cdot {h}_{t}\)). These terms are particularly important for capturing idiosyncratic spikes in mortality that result from nation-wide conditions, such as influenza outbreaks, as well as any systematic changes in the accounting methodology of mortality by the CDC. Comparisons of Extended Data Fig. 4b,c illustrates how the inclusion of these terms in the model alters the ability of the model to capture unusual spikes in mortality that are not captured by other model elements, including the trends listed above, TCs, and temperature.

    Overall, the fit for this model is high (in-sample adjusted R2 = 0.93 with 25,062 degrees of freedom). Extended Data Fig. 3 overlays predictions with observations for all states (same as in Fig. 1c). We find that all of the non-parametric controls listed above are important for passing standard specification checks. For example, failure to account for trends flexibly enough causes estimated leads to deviate from zero or randomization-based placebo tests (described below) to recover non-zero central estimates. These results are unchanged if we use a Poisson regression specification (Extended Data Fig. 6a).

    In a robustness test, we interact the month-of-sample fixed effects with 3 region indicator variations. We continue to obtain our main findings after introducing these additional 2,062 parameters to the model, although the estimates become much noisier and attenuate slightly. Both of these effects are well understood results of including a large number of highly flexible variables that absorb a meaningful fraction of the true variation in the independent variable69.

    We evaluate the distribution of the unmodelled variation represented by the error term ϵit and find that it essentially follows a Normal distribution except with slightly positive kurtosis (Supplementary Fig. 5a). The distribution of these residuals appears stationary throughout the sample period and independent over time (Supplementary Fig. 5b). The consistency of the distribution of these errors is attributed to the high degree of flexibility in the non-parametric terms of our econometric specification, which are able to capture those components of the data-generating process that would otherwise appear as auto-correlated errors. On the basis of this evaluation, we construct OLS standard-error estimates as the underlying assumptions for these estimates appear to be reasonably satisfied. In addition, we find strong support for this modelling choice when we conduct a variety of permutation tests for statistical significance (Extended Data Fig. 5), all of which indicate that our asymptotic estimates for confidence intervals are correctly (and possibly conservatively) sized and our tests for statistical significance correctly powered. Notably, these permutation tests do not rely on the assumptions used to estimate these confidence intervals, thus they can be considered independent corroboration for the validity of this approach.

    Cumulative effects

    To compute the total effect after the TC makes landfall, we estimate the cumulative sum of β for each  [−72, 240]. We compute \({\Omega }_{{\ell }}={\sum }_{k=o}^{{\ell }}{\beta }_{k}\), which denotes the cumulative impact of an additional 1 ms−1 wind speed incidence on mortality months after a TC event. We account for the estimated covariance of β when estimating uncertainty in Ω. We normalize the sums relative to the impact one month prior to the TC,  = −1 such that Ω−1 = 0.

    Randomization-based placebo tests

    Given the complexity of our model, the long delays we study, and the absence of prior analyses of long-run total mortality from TCs, it is not possible to subjectively evaluate our econometric analysis against any prior benchmark. In such a context, there is risk of unknowingly recovering a spurious estimate generated as an artefact of our model specification. A strong test designed to avoid such artefacts is to ensure that model estimates of TC impacts on mortality are unbiased in a variety of situations where the structure of the association has been manipulated. In four tests, we shuffle the true TC data in different ways. In each case, this shuffling should break any correlation between TC incidence and mortality such that an unbiased estimate of the effect of shuffled TCs on mortality is zero. However, in each case, some of the structure in the original TC data are allowed to remain in the shuffled TC data. For example, randomization within a state over time retains the average cross-sectional patterns of TC incidence, but destroys any time- series structure. Thus, these tests allow us to examine whether, in each case, the remaining structure generates artefacts in the model that would produce a spurious result, also known as a negative exposure control70. Any non-zero correlation, on average, would indicate a biased model where the bias is driven by the non-randomized components of the original TC data.

    Within each type of randomization we scramble TC assignment 1,000 times and run the linear version of the model (equation (8)) on each re-sampled version of the data. Our four randomizations are illustrated graphically in Supplementary Fig. 7 and described below:

    • Total randomization shuffles TC events across all state-by-month observations. This tests whether the unconditional marginal distribution of TC events, which has a long right tail, could generate bias. Results are shown by light blue boxes in Fig. 1g.

    • Within-state randomization shuffles the sequencing of TCs that a state experiences over time. TCs are always assigned to the correct state, but the month and year assigned to each storm is random. The cross-sectional average pattern of storm incidence is preserved in the data. Thus, this tests whether time-invariant cross-sectional patterns across states generate spurious correlations. Results are shown by dark blue boxes in Fig. 1g.

    • Within-month randomization shuffles the TC incidence across states within each month-of-sample. TCs are always assigned to the correct month and year, but the state assigned to each storm is random. The average time-series structure of TC incidence nation-wide is preserved. Thus, this tests whether national or seasonal trends, which are nonlinear, could bias estimates produced by this model. Results are shown by maroon boxes in Fig. 1g.

    • Across-state shuffles complete TC times-series across states, keeping the timing and sequence of storms correct as blocks. TCs are always assigned to the correct month and year, and the sequence of storms experienced by a state is always a continuous sequence that is observed in the data. However, the state that is assigned that sequence is randomly chosen. This tests whether trends within a state and within the sequence of storms that a state experiences could generate bias. This test differs from the within month randomization because state-level trends often differ across states (see Extended Data Fig. 1 and Extended Data Fig. 3) and there are complex seasonal patterns that could potentially affect estimates. Results are shown by red boxes in Fig. 1g.

    The estimated impact of TCs in each of these placebo tests is zero on average, to within a high degree of precision. Extended Data Fig. 5 illustrates distributions of estimates for all lags. These results demonstrate that non-exchangeability across states within a month, across months within a state, or across states (conditional on month of sample) does not confound our analysis; indicating that the rich set of fixed effects and trends successfully adjust for many patterns of TC incidence and/or mortality such that the remaining conditional variation is as good as random.

    Permutation tests for statistical significance

    In addition to establishing the unbiasedness of our main point estimates, the four randomizations above can be utilized to serve a second function: estimating statistical significance of our estimates. These randomizations enable approximate permutation tests71, allowing for different types of autocorrelation to remain in the TC data. We use these randomizations to examine the likelihood of randomly obtaining an entire impulse-response function similar our actual estimate, if in reality no such relationship exists in the data. To do this, we jointly test the significance of all true cumulative estimates Ω against the null hypothesis that a similarly extreme sequence of estimates is generated randomly. Extended Data Fig. 5 overlays the true estimates for Ω on distributions of similar estimates from each randomization. P values for individual lag terms \(\left({p}_{{\ell }}=\Pr \left(\left|{\Omega }_{{\ell }}^{{\rm{randomized}}}\right| > \left|{\Omega }_{{\ell }}\right|\right)\right)\) are plotted in the right subpanels and are all individually statistically significant (P < 0.05) for  < 150 months in each randomization. However, the significance of the complete sequence of coefficients that together compose the entire impulse-response function is far greater. We compute a joint P value for the full impulse response between 0 and 172 months \(({p}_{{\ell }}=\Pr ({\cap }_{{\ell }=0}^{172}(| {\Omega }_{{\ell }}^{{\rm{randomized}}}| > | {\Omega }_{{\ell }}| ))\) that ranges from P = 0.012 to P = 0.0012 across the randomization approaches (Extended Data Fig. 5). We conclude that it would be extremely unlikely to obtain an impulse-response function as extreme as our main result due to chance.

    Subsamples by age, race and cause of death

    In addition to the all-cause mortality rate for the entire population, we also present the results stratified by age, race, and cause of death. For the six cause-specific mortality rates we compute mortality per 100,000 of the total population in that state in the time period (for example, total number of deaths from cardiovascular disease divided by total population times 100,000). ‘Other’ mortality is the difference between the total deaths and the sum of all the other causes. For age groups and race we report mortality risk as the outcome of interest. For example, for the Black population, we construct the mortality rate for Black people as the number of deaths of Black people in the state divided by the Black population. We do the same procedure by age group. We also report the mortality by these strata as a proportion of the total deaths traceable to TC incidence, see Extended Data Fig. 7.

    Subsamples by average TC risk

    To evaluate whether there is heterogeneity in the mortality response of states that are frequently exposed to TCs compared to those infrequently exposed, we stratify the sample by the average TC incidence they experience. We allow the mortality impulse response to differ based on quartiles of states, sorted by their average TC incidence. We implement this by including and interaction with an indicator variable for the quartile of their average wind speed incidence, following the general approach for modelling adaptation developed in refs. 29,33. Average wind incidence is a measure of the expected TC risk a population bears, which informs preventive risk reduction investments, behaviours, or other adaptive actions they take to reduce the expected harm from TCs. We approximate this measure by computing as the mean wind speed in each state i across the period t in our sample

    $$wind{\rm{\_}}{speed}_{i}=\frac{{\sum }_{t=1}^{1,032}wind{\rm{\_}}{speed}_{it}}{\mathrm{1,032}}$$

    and assigning the quartiles of these means to each state. We estimate a model that allows each quartile to express a different impulse response to TCs and observe little difference in the impact of TCs on mortality between the second through fourth quartile (Extended Data Fig. 6c). The effect for the second through fourth quartile are not statistically significantly different than the effect for the second through fourth quartile, combined (P = 0.38). Thus, to improve the efficiency of our model and limit unnecessary noise in our estimates, we pool quartiles 2–4 to create the ‘high average incidence’ group in the main results (shown in Fig. 2f) (‘high incidence’ in equation (9)). The ‘low average incidence’ group is the first quartile of average wind speed alone (‘low incidence’ in equation (9)). We additionally evaluate whether the spatial distribution of populations relative to the coast, within states, alters the mortality response to TCs. Stratifying states on the basis of the average fraction of the population that lives in coastal counties, we fail to find evidence that states with high concentrations of coastal populations are systematically different from states with little or no coastal population (Extended Data Fig. 6d). Lastly, we evaluate whether the overall spatial correlation between populations and average wind speed incidence, within each state, alters the mortality response to TCs. Stratifying states on the basis of within-state spatial correlation across 0.125° × 0.125° pixels, we fail to find evidence that states with higher spatial correlations are systematically different from states with little or negative correlations (Extended Data Fig. 6e).

    Nonlinear effects of TCs

    We evaluate whether the mortality impact of a TC is nonlinear in the physical intensity of the event. This could occur, for example, if more extreme TC events generate exponentially more physical damage51,72 or if they elicit different government responses18,73. Empirically, we find that excess mortality 180 months after a TC is well approximated by a linear function of max wind incidence, particularly for TCs with area-average max wind speeds between 0 and 20 ms−1, which is the majority of events (93%) in our sample (Extended Data Fig. 8). However, for the most extreme events (>30 ms−1, 1.4% of events) we find that excess mortality is generally lower than a linear function would predict, although these nonlinear effects are not themselves statistically significant. We lack the data to fully evaluate the underlying causes of this nonlinearity, but believe it is an important topic for future study. For example, it is possible that societal responses to the most extreme events (for example, disaster relief) are more effective at alleviating mortality impacts of TCs because these events attract a disproportionate quantity of attention, compared to less extreme events that are also harmful but less salient74. Regardless of their cause, we account for these non-linearities in calculations below as they contain information on how populations in CONUS have adapted to their TC climates5,73.

    To estimate nonlinear effects of TCs, we estimate a model that is identical to the benchmark linear model in equation (1), but it allows the magnitude of the TC mortality impulse-response function to be cubic in TC incidence. The motivation for this approach is the possibility that the relationships between wind speed and long-run mortality does not increase linearly. For example, very high wind speeds may cause extreme damages and/or elicit greater governmental and humanitarian responses, which would mean that a unit of increase from 40 to 41 ms−1 wind speed may have a larger or small mortality impact compared to an increase from 5 to 6 ms−1. Since the nonlinear impact of TCs may be influenced by the historical TC experience and baseline TC risk of populations, this model also allows for the nonlinear response to differ based on the risk categorization for each state:

    $$\begin{array}{l}mortality{\rm{\_}}{rate}_{it}=\\ \,\mathop{\sum }\limits_{{\ell }=-72}^{240}\mathop{\sum }\limits_{r=1}^{3}({\theta }_{r,{\ell }}^{low{\rm{\_}}incidence}\cdot wind{\rm{\_}}{speed}_{i,t-{\ell }}^{r}\cdot {Q}_{i}^{low{\rm{\_}}incidence}\\ \,+\,{\theta }_{r,{\ell }}^{high{\rm{\_}}incidence}\cdot wind{\rm{\_}}{speed}_{i,t-{\ell }}^{r}\cdot {Q}_{i}^{high{\rm{\_}}incidence})\\ \,+\,controls+{{\epsilon }}_{it}\end{array}$$

    (9)

    where r indicates an exponent and the controls are identical to those in equation (8). \({Q}_{i}^{low{\rm{\_}}incidence}\) (\({Q}_{i}^{high{\rm{\_}}incidence}\)) is an indicator variable that is set to one if state i is in the low incidence (high incidence) group. The coefficients θ1, θ2 and θ3 separately capture the cubic relationship between wind speed incidence and mortality for low and high-risk states in each lag period. Extended Data Fig. 8 displays the cumulative impact estimated using both the linear and nonlinear models after 180 months. These results are unchanged if we alternatively use a cubic spline regression specification (Extended Data Fig. 8). The low-risk response is slightly convex relative to the linear estimate, while the high-risk response is slightly concave. In both cases, impacts are relatively well approximated by the linear version of the model and only diverge (insignificantly) at very high levels of incidence that are rare in sample. Distributions for in sample frequency are shown in lower panels of Extended Data Fig. 8.

    Computing mortality burdens

    The total impacts of all TCs on mortality are estimated using each version of the model, presented in Supplementary Table 1. We compute the excess mortality from TCs by state, month, and TC. These estimates are presented in Figs. 3 and 4 and Supplementary Tables 1 and 2.

    Figure 4 displays the estimated excess mortality from TCs by state, age and race, computed using equation (8) applied to equation (7). Figure 4a presents our estimated full TC mortality burden, similar to equation (7) but by state (mortality_burdenit) as an average proportion of total deaths (mortalityit) in each state between 1950 and 2015:

    $${proportion}_{i}=\frac{{\sum }_{t\in {\rm{m}}{\rm{o}}{\rm{n}}{\rm{t}}{\rm{h}}}mortality{\rm{\_}}{burden}_{it}}{{\sum }_{t\in {\rm{m}}{\rm{o}}{\rm{n}}{\rm{t}}{\rm{h}}}{mortality}_{it}}$$

    Similarly, we estimate the proportion by state and age group (proportioni,a), shown in Fig. 4d,f. Proportion and total excess mortality for the Black population is based on mortality burden estimated with the mortality risk for the Black population, therefore \({proportion}_{Black}=\frac{{\sum }_{t\in {\rm{m}}{\rm{o}}{\rm{n}}{\rm{t}}{\rm{h}}}{\sum }_{i\in {\rm{s}}{\rm{t}}{\rm{a}}{\rm{t}}{\rm{e}}}mortality{\rm{\_}}{burden}_{it,{\rm{B}}{\rm{l}}{\rm{a}}{\rm{c}}{\rm{k}}}}{{\sum }_{t\in {\rm{m}}{\rm{o}}{\rm{n}}{\rm{t}}{\rm{h}}}{\sum }_{i\in {\rm{s}}{\rm{t}}{\rm{a}}{\rm{t}}{\rm{e}}}{mortality}_{it,{\rm{B}}{\rm{l}}{\rm{a}}{\rm{c}}{\rm{k}}}}\).

    Figure 4g illustrates the impact of the TC climate on the geographic differences in average annual all-cause mortality rate between states that do not experience TCs (‘non-TC states’) and states that do (‘TC states (actual)’), in this context. We also subtract the average annual TC mortality burden from the actual average annual mortality for each TC-impacted state (‘TC states without TCs’).

    Decomposing trends in mortality burden

    We examine the differences in TC events and population distribution before 2001 and after 2001 in order to understand why the mortality burden after 2001 is increasing more rapidly than it did before 2001 (Fig. 4h–j). Figure 4h shows the distribution of the number of TCs that made landfall each year before 2001 and after 2001. Similarly, Fig. 4i shows the maximum annual wind speed per year and Fig. 4j plots the average wind speed per year as experienced by a proportion of the CONUS population. The changes in the distribution of TC events affecting CONUS after 2001 were themselves probably caused by a combination of factors, including warmer sea surface temperatures11,75 and reductions of anthropogenic aerosol emissions76,77 (which create an environment more amenable to TC intensification); and shifts in steering winds78 (which direct a larger fraction of TCs to landfall in CONUS after formation). We note that identifying factors driving the TC climate remains an active area of research56,79.

    To understand the 1950 to 2015 trend in the national aggregate mortality burden, we re-estimate mortality_burdent for each month of the sample with different populations to decompose the long-term trend based on various population patterns. We first replace populationi,t+ from equation (7) with fixed 1950 or 2015 populations (Fig. 4k, red and maroon lines). To generate the yellow line in Fig. 4k, we replace populationi,t+ with an estimate of the 2015 population ‘deflated’ to 1950 levels. Specifically, we first compute a national population deflation fraction,

    $$\Delta =\frac{{\sum }_{i\in {\rm{s}}{\rm{t}}{\rm{a}}{\rm{t}}{\rm{e}}}{population}_{i,2015}-{\sum }_{i\in {\rm{s}}{\rm{t}}{\rm{a}}{\rm{t}}{\rm{e}}}{population}_{i,1950}}{{\sum }_{i\in {\rm{s}}{\rm{t}}{\rm{a}}{\rm{t}}{\rm{e}}}{population}_{i,1950}}$$

    where populationi,2015 is the state-specific population in 2015 and populationi,1950 is the state population in 1950. We then calculate

    $$deflated\_\,{population}_{i,2015}=\frac{{\sum }_{i\in {\rm{s}}{\rm{t}}{\rm{a}}{\rm{t}}{\rm{e}}}{population}_{i,2015}}{1+\Delta }$$

    and apply deflated_populationi,2015 to equation (7). This value is an adjusted state-level population that allows the total national population to match 1950 level but have a relative spatial distribution that reflects 2015.

    Reporting summary

    Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.

    [ad_2]

    Source link

  • Trumbore, S. E. in Quaternary Geochronology: Methods and Applications (eds Stratton Noller, J. et al.) 41–60 (AGU, 2000).

  • Karkanas, P. & Goldberg, P. Reconstructing Archaeological Sites: Understanding the Geoarchaeological Matrix (Wiley, 2018).

  • Bailey, G. Time perspectives, palimpsests and the archaeology of time. J. Anthropol. Archaeol. 26, 198–223 (2007).

    Article 

    Google Scholar
     

  • Mallol, C. & Hernández, C. M. Advances in palimpsest dissection. Quat. Int. 417, 1–2 (2016).

    Article 

    Google Scholar
     

  • Binford, L. R. Willow smoke and dogs’ tails: hunter-gatherer settlement systems and archaeological site formation. Am. Antiq. 45, 4–20 (1980).

    Article 

    Google Scholar
     

  • Kelly, R. L. Hunther-gatherer mobility strategies. J. Anthropol. Res. 39, 277–306 (1983).

    Article 

    Google Scholar
     

  • Bettinger, R. L., Garvey, R. & Tushingham, S. Hunter-Gatherers: Archaeological and Evolutionary Theory (Springer, 2015).

  • Hamilton, M. J., Milne, B. T., Walker, R. S. & Brown, J. H. Nonlinear scaling of space use in human hunter-gatherers. Proc. Natl Acad. Sci. USA 104, 4765–4769 (2007).

    Article 
    ADS 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Hamilton, M. J., Milne, B. T., Walker, R. S., Burger, O. & Brown, J. H. The complex structure of hunter-gatherer social networks. Proc. R. Soc. B 274, 2195–2203 (2007).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Draper, P. Crowding among hunter-gatherers: The !Kung bushmen. Science 182, 301–303 (1973).

    Article 
    ADS 
    CAS 
    PubMed 

    Google Scholar
     

  • Ember, C. R. Residential variation among hunter-gatherers. Behav. Sci. Res. 10, 199–227 (1975).

    Article 

    Google Scholar
     

  • Winterhalder, B., Baillargeon, W., Cappelletto, F., Daniel, I. R. Jr & Prescott, C. The population ecology of hunter-gatherers and their prey. J. Anthropol. Archaeol. 7, 289–328 (1988).

    Article 

    Google Scholar
     

  • Vandevelde, S., Brochier, J. E., Petit, C. & Slimak, L. Establishment of occupation chronicles in Grotte Mandrin using sooted concretions: rethinking the Middle to Upper Paleolithic transition. J. Hum. Evol. 112, 70–78 (2017).

    Article 
    PubMed 

    Google Scholar
     

  • Vandevelde, S. et al. Identification du rythme annuel de précipitation des carbonates pariétaux pour un calage micro-chronologique des occupations archéologiques pyrogéniques: cas de la Grotte Mandrin (Malataverne, Drôme, France). BSGF Earth Sci. Bull. 192, 1–22 (2021).

    Article 

    Google Scholar
     

  • Slimak, L. et al. Modern human incursion into Neanderthal territories 54,000 years ago at Mandrin, France. Sci. Adv. 8, eabj9496 (2022).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Lugli, F. et al. Tracing the mobility of a Late Epigravettian (~13 ka) male infant from Grotte di Pradis (Northeastern Italian Prealps) at high-temporal resolution. Sci. Rep. 12, 8104 (2022).

  • Galván, B. et al. New evidence of early Neanderthal disappearance in the Iberian Peninsula. J. Hum. Evol. 75, 16–27 (2014).

    Article 
    PubMed 

    Google Scholar
     

  • Garralda, M. D. et al. Neanderthals from El Salt (Alcoy, Spain) in the context of the latest Middle Palaeolithic populations from the southeast of the Iberian Peninsula. J. Hum. Evol. 75, 1–15 (2014).

    Article 
    PubMed 

    Google Scholar
     

  • Mallol, C. et al. The black layer of Middle Palaeolithic combustion structures. Interpretation and archaeostratigraphic implications. J. Archaeolog. Sci. 40, 2515–2537 (2013).

    Article 

    Google Scholar
     

  • Leierer, L. et al. Insights into the timing, intensity and natural setting of Neanderthal occupation from the geoarchaeological study of combustion structures: a micromorphological and biomarker investigation of El Salt, unit xb, Alcoy, Spain. PLoS ONE 14, e0214955 (2019).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Mayor, A., Hernández, C. M., Machado, J., Mallol, C. & Galván, B. On identifying Palaeolithic single occupation episodes: archaeostratigraphic and technological approaches to the Neanderthal lithic record of stratigraphic unit xa of El Salt (Alcoi, eastern Iberia). Archaeol. Anthropol. Sci. 12, 84 (2020).

    Article 

    Google Scholar
     

  • Machado, J., Molina, F. J., Hernández, C. M., Tarriño, A. & Galván, B. Using lithic assemblage formation to approach Middle Palaeolithic settlement dynamics: El Salt Stratigraphic Unit x (Alicante, Spain). Archaeol. Anthropol. Sci. 9, 1715–1743 (2017).

    Article 

    Google Scholar
     

  • Sternberg, R. & Lass, E. H. E. in Kebara Cave, Mt. Carmel, Israel. The Middle and Upper Palaeolithic Archaeology. Part I (eds Bar-Yosef, O. & Meignen, L.) 123–130 (Peabody Museum of Archaeologyand Ethnology, Harvard University, 2007).

  • Carrancho, Á., Villalaín, J. J., Vallverdú, J. & Carbonell, E. Is it possible to identify temporal differences among combustion features in Middle Palaeolithic palimpsests? The archaeomagnetic evidence: a case study from level O at the Abric Romaní rock-shelter (Capellades, Spain). Quat. Int. 417, 39–50 (2016).

    Article 

    Google Scholar
     

  • Zeigen, C., Shaar, R., Ebert, Y. & Hovers, E. Archaeomagnetism of burnt cherts and hearths from Middle Palaeolithic Amud Cave, Israel: tools for reconstructing site formation processes and occupation history. J. Archaeolog. Sci. 107, 71–86 (2019).

    Article 

    Google Scholar
     

  • Molina-Cardín, A. et al. Updated Iberian Archeomagnetic Catalogue: new full vector paleosecular variation curve for the last three millennia. Geochem. Geophys. Geosyst. 19, 3637–3656 (2018).

    Article 
    ADS 

    Google Scholar
     

  • Pavón Carrasco, F. J., Osete, M. L., Torta, J. M. & De Santis, A. A geomagnetic field model for the Holocene based on archaeomagnetic and lava flow data. Earth Planet. Sci. Lett. 388, 98–109 (2014).

    Article 
    ADS 

    Google Scholar
     

  • Schanner, M., Korte, M.- & Holschneider, M. ArchKalmag14k: a Kalman–Filter based global geomagnetic model for the Holocene. J. Geophys. Res. 127, e2021JB023166 (2022).

    Article 
    ADS 

    Google Scholar
     

  • Constable, C., Korte, M. & Panovska, S. Persistent high paleosecular variation activity in southern hemisphere for at least 10000 years. Earth Planet. Sci. Lett. 453, 78–86 (2016).

    Article 
    ADS 
    CAS 

    Google Scholar
     

  • Osete, M. L. et al. Two archaeomagnetic intensity maxima and rapid directional variation rates during the Early Iron Age observed at Iberian coordinates. Implications on the evolution of the Levantine Iron Age Anomaly. Earth Planet. Sci. Lett. 533, 116047 (2020).

    Article 
    CAS 

    Google Scholar
     

  • Herrejón Lagunilla, Á., Carrancho, Á., Villalaín, J. J., Mallol, C. & Hernández, C. M. An experimental approach to the preservation potential of magnetic signatures in anthropogenic fires. PLoS ONE 14, e0221592 (2019).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Sistiaga, A., Mallol, C., Galván, B. & Summons, R. E. The Neanderthal meal: a new perspective using faecal biomarkers. PLoS ONE 9, e101045 (2014).

    Article 
    ADS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Leierer, L. et al. It’s getting hot in here—microcontextual study of a potential pit hearth at the Middle Paleolithic site of El Salt, Spain. J. Archaeolog. Sci. 123, 105237 (2020).

    Article 
    CAS 

    Google Scholar
     

  • Galván, B. El Salt (Alcoi, Alicante): estado actual de las investigaciones. Recerques del Museu D’Alcoi 1, 73–80 (1992).


    Google Scholar
     

  • Fumanal García, M. P. El yacimiento musteriense de El Salt (Alcoi, País Valenciano). Rasgos geomorfológicos y climatoestratigrafía de sus registros. SAGVNTVM 27, 39–55 (1994).


    Google Scholar
     

  • Galván, B. et al. in Pleistocene and Holocene Hunter-Gatherers in Iberia and the Gibraltar Strait: The Current Archaeological Record (ed. Sala Ramos, R.) 380–388 (Fundación Atapuerca, Servicio de Publicaciones de la Universidad de Burgos, 2014).

  • Machado, J. & Pérez, L. Temporal frameworks to approach human behaviour concealed in Middle Palaeolithic palimpsests: a high-resolution example from El Salt Stratigraphic Unit x (Alicante, Spain). Quat. Int. 417, 66–81 (2016).

    Article 

    Google Scholar
     

  • Dunlop, D. J. & Özdemir, Ö. Rock Magnetism. Fundaments and Frontiers (Cambridge Univ. Press, 1997).

  • Carrancho, Á. & Villalaín, J. J. Different mechanism of magnetisation recorded in experimental fires: archaeomagnetic implications. Earth Planet. Sci. Lett. 312, 176–187 (2011).

    Article 
    ADS 
    CAS 

    Google Scholar
     

  • Panovska, S., Constable, C. G. & Korte, M. Extending global continuous geomagnetic field reconstructions on timescales beyond human civilization. Geochem. Geophys. Geosyst. 19, 4757–4772 (2018).

    Article 
    ADS 

    Google Scholar
     

  • Fisher, R. Dispersion on a sphere. Proc. R. Soc. Lond. A 217, 295–305 (1953).

    Article 
    ADS 
    MathSciNet 

    Google Scholar
     

  • Wessel, P. et al. The Generic Mapping Tools version 6. Geochem. Geophys. Geosyst. 20, 5556–5564 (2019).

    Article 
    ADS 

    Google Scholar
     

  • Carrancho, Á., Villalaín, J. J., Vergès, J. M. & Vallverdú, J. Assessing post-depositional processes in archaeological cave fires through the analysis of archaeomagnetic vectors. Quat. Int. 275, 14–22 (2012).

    Article 

    Google Scholar
     

  • Chadima, M. & Hrouda, F. Remasoft 3.0—a user-friendly paleomagnetic data browser and analyzer. Travaux Géophysiques XXVII, 20–21 (2006).


    Google Scholar
     

  • Leonhardt, R. Analyzing rock magnetic measurements: the RockMagAnalyzer 1.0 software. Comput. Geosci. 32, 1420–1431 (2006).

    Article 
    ADS 

    Google Scholar
     

  • Bargalló, A., Gabucio, M. J. & Rivals, F. Puzzling out a palimpsest: testing an interdisciplinary study in level O of Abric Romaní. Quat. Int. 417, 51–65 (2016).

    Article 

    Google Scholar
     

  • Machado, J., Hernández, C. M., Mallol, C. & Galván, B. Lithic production, site formation and Middle Palaeolithic palimpsest analysis: in search of human occupation episodes at Abric del Pastor stratigraphic unit IV (Alicante, Spain). J. Archaeolog. Sci. 40, 2254–2273 (2013).

    Article 

    Google Scholar
     

  • Machado, J., Mayor, A., Hernández, C. M. & Galván, B. Lithic refitting and the analysis of Middle Palaeolithic settlement dynamics: a high-temporal resolution example from El Pastor rock shelter (eastern Iberia). Archaeolog. Anthropol. Sci. 11, 4539–4554 (2019).

    Article 

    Google Scholar
     

  • Spagnolo, V. et al. Climbing the time to see Neanderthal behaviour’s continuity and discontinuity: SU 11 of the Oscurusciuto rockshelter (Ginosa, southern Italy). Archaeolog. Anthropol. Sci. 12, 1–30 (2020).


    Google Scholar
     

  • PaleomagUCM/El-Salt: v1.0 (v1.0). Zenodo https://doi.org/10.5281/zenodo.10931465 (2024).

[ad_2]

Source link

  • Companies inadvertently fund online misinformation despite consumer backlash

    [ad_1]

    Background on digital advertising

    The predominant business model of several mainstream digital media platforms relies on monetizing attention via advertising3. While these platforms typically offer free content and services to individual consumers, they generate revenue by serving as an intermediary or advertising exchange connecting advertisers with independent websites that want to host advertisements. To do so, platforms run online auctions to algorithmically distribute advertising across websites, known as ‘programmatic advertising’. For example, Google distributes advertising in this manner to more than two million non-Google sites that are part of the Google Display Network. This allows websites to generate revenue for hosting advertising, and they share a percentage of this payment with the platform. In the USA, more than 80% of digital display advertisements are placed programmatically16. We refer to these advertising exchanges as digital advertising platforms and use the term digital platforms to collectively refer to all the services offered by such media platforms.

    We examine the role of advertising companies and digital advertising platforms in monetizing online misinformation. While in other forms of (offline) media, advertisers typically have substantial control over where their advertisements appear, advertising placement through digital advertising platforms is mainly automated. Since most companies do not have the capacity to participate in high-frequency advertising auctions that require them to place individual bids for each advertising slot they are interested in, they typically outsource the bidding process to an advertising platform. Such programmatic advertising gives companies relatively less control over where their advertisements end up online. However, companies can take steps to reduce advertising on misinformation websites, such as by only being part of advertising auctions for a select list of credible websites or blocking advertisements from appearing on specific misinformation outlets.

    Collecting website data

    We collect data on misinformation websites in three steps. First, we use a dataset maintained by NewsGuard. This company rates all the news and information websites that account for 95% of online engagement in each of the five countries where it operates. Journalists and experienced editors manually generate these ratings by reviewing news and information websites according to nine apolitical journalistic criteria. Recent research has used this dataset to identify misinformation websites6,66,67. In this paper, we consider each website that NewsGuard rates as repeatedly publishing false content between 2019 and 2021 to be a misinformation website and all others to be non-misinformation websites, leading to a set of 1,546 misinformation websites and 6,499 non-misinformation websites. To get coverage throughout our study period, we sample websites provided by NewsGuard from the start, middle and end of each year from 2019 to 2021. Additionally, we also sample websites from January 2022 and June 2022 to account for websites that may have existed during our study period and discovered later. Supplementary Table 3 summarizes the characteristics of this dataset. Our NewsGuard dataset contains websites across the political spectrum, including left-leaning websites (for example, https://www.palmerreport.com/ and https://occupydemocrats.com/), politically neutral websites (for example, https://rt.com/ and https://www.nationalenquirer.com), and right-leaning websites (for example, https://www.thegatewaypundit.com/ and http://theconservativetreehouse.com/).

    Note that prior research that has used the NewsGuard dataset has often used the term ‘untrustworthy’ to describe websites6,67. Such research has used NewsGuard’s aggregate classification whereby a site that scores below a certain threshold (60 points) on NewsGuard’s weighted score system is labelled as untrustworthy. Instead of using NewsGuard’s overall score for a website, we use the first criterion classified by NewsGuard for each website—that is, whether a website repeatedly publishes false news to identify a set of 1,546 misinformation websites. While 94% of the NewsGuard misinformation websites we identify in this manner are also untrustworthy based on NewsGuard’s classification, only about 52% of the untrustworthy websites are misinformation websites or websites that repeatedly publish false news. Our measure of misinformation is, therefore, more conservative than prior work using NewsGuard’s ‘untrustworthy’ label.

    In addition to the NewsGuard dataset, we use a list of websites provided by the GDI. This non-profit organization identifies disinformation by analysing both the content and context of a message, and how they are spread through networks and across platforms68. In this way, GDI maintains a list of monthly-updated websites, which it also shares with interested advertising tech platforms to help reduce advertising on misinformation websites. The GDI list allows us to identify 1,869 additional misinformation websites. Finally, we augment our list of misinformation websites with 396 additional ones used in prior work69,70. Among the websites that NewsGuard rated as non-misinformation (at any point in our sample), 310 websites were considered to be misinformation websites by our other sources or by NewsGuard itself (during a different period in our sample). We categorize these websites as misinformation websites given their risk of producing misinformation.

    Altogether, our website dataset consists of 10,310 websites, including 3,811 misinformation and 6,499 non-misinformation websites. Similar to prior work6,67, our final measure of misinformation is at the level of the website or online news outlet. Aggregating article-level information and using website-level metadata is meaningful since it reduces noise when arriving at a website-level measure. Finally, we use data from SEMRush, a leading online analytics platform, to determine the level of monthly traffic received by each website from 2019 to 2021.

    Consumer experiment design

    This study was reviewed by the Stanford University Institutional Review Board (Protocol No. IRB-63897) and the Carnegie Mellon University Institutional Review Board (protocol no. IRB00000603). Our study was pre-registered at the American Economic Association’s Registry under AEARCTR-0009973. Informed consent was obtained from all participants at the beginning of the survey.

    Setting and sample recruitment

    We recruited a sample of US internet users via CloudResearch. CloudResearch screened respondents for our study so that they are representative of the US population in terms of age, gender and race based on the US Census (2020). It is important to note that while we recruited our sample to be representative on these dimensions to improve the generalizability and external validity of our results, our sample is a diverse sample of US internet users, which is not necessarily representative of the US population on other dimensions71. To ensure data quality, we include a screener in our survey to check whether participants pay attention to the information provided. Only participants who pass this screener can proceed with the survey. Our total sample includes 4,039 participants, who are randomized into five groups approximately evenly.

    The flow of the survey study is shown in Supplementary Fig. 1. We begin by asking participants to report demographics such as age, gender and residence. From a list of trustworthy and misinformation outlets, we then ask participants questions about their behaviours in terms of the news outlets they have used in the past 12 months, their trust in the media (on a 5-point scale), the online services or platforms they have used and the number of petitions they have signed in the past 12 months.

    Initial gift card preferences

    We then inform participants that one in five (that is, 20% of all respondents) who complete the survey will be offered a US$25 gift card from a company of their choice out of six company options. Respondents are asked to rank the six gift card companies on a scale from their first choice (most preferred) to their sixth choice (least preferred). These six companies belong to one of three categories: fast food, food delivery and ride-sharing. All six companies appeared on the misinformation websites in our sample during the past three years (2019–2021), offer items below US$25, and are commonly used throughout the USA. The order in which the six companies are presented is randomized at the respondent level. As a robustness check, we also ask respondents to assign weights to each of the six gift card options. This question gives respondents greater flexibility by allowing them to indicate the possibility of indifference (that is, equal weights) between any set of options. We then ask participants to confirm which gift card they would like to receive if they are selected to ensure they have consistent preferences regardless of how the question is asked. At this initial elicitation stage, the respondents did not know that they will get another chance to revise their choice. Hence, these choices can be thought of as capturing their revealed preference.

    Information treatments

    All participants in the experiment are given baseline information on misinformation and advertising. This is meant to ensure that all participants in our experiment are made aware of how we define misinformation along with examples of a few misinformation websites (including right-wing, neutral and left-wing misinformation websites), how misinformation websites are identified, and how companies advertise on misinformation websites (via an illustrative example) and use digital platforms to automate placing advertisements.

    Participants are then randomized into one control and four treatment groups, in which the information treatments are all based on factual information from our data and prior research. We use an active control design to isolate the effect of providing information relevant to the practice of specific companies on people’s behaviour9. Participants in the control group are given generic information based on prior research that is unrelated to advertising companies or platforms but relevant to topic of news and misinformation.

    In our first ‘company only’ treatment group (T1), participants are given factual information stating that advertisements from their top choice gift card company appeared on misinformation websites in the recent past. Based on their preferences, people may change their final gift card preference away from their initial top-ranked company after receiving this information. It is unclear, however, whether advertising on misinformation websites would cause a sufficient change in consumption patterns and which sets of participants may be more affected.

    Our second ‘platform only’ treatment group (T2) informs participants that companies using digital advertising platforms were about 10 times more likely to appear on misinformation websites than companies that did not use such platforms in the recent past. This information treatment measures the effects of digital advertising platforms in financing misinformation news outlets. Since it does not contain information about advertising companies, it practically serves as a second control group for our company-level outcome and aims to measure how people may respond to our platform-related outcome.

    Because our descriptive data suggest that the use of digital advertising platforms amplifies advertising revenue for misinformation outlets, we are interested in measuring how consumers respond to a specific advertising company appearing on misinformation websites when also informed of the potential role of digital advertising platforms in placing companies’ advertising on misinformation websites. It is unclear whether consumers will attribute more blame to companies or advertising platforms for financing misinformation websites when informed about the role of the different stakeholders in this ecosystem. For this reason, our third ‘company and platform’ treatment (T3) combines information from our first two treatments (T1 and T2). Similar to T1, participants are given factual information that advertisements from their top choice gift card company appeared on misinformation websites in the recent past. Additionally, we informed participants that their top choice company used digital advertising platforms and companies that used such platforms were about ten times more likely to appear on misinformation websites than companies that did not use digital advertising platforms, as mentioned in T2.

    Finally, since several advertising companies appear on misinformation websites, we would like to determine whether informing consumers about other advertising companies also appearing on misinformation websites changes their response towards their top choice company. In our fourth company-ranking treatment (T4), participants are given factual information, which states that “In the recent past, ads from all six companies below repeatedly appeared on misinformation websites in the following order of intensity”, and provided with a ranking from one of three years in our study period—that is, 2019, 2020 or 2021. We personalize these rankings by providing truthful information based on data from different years in the recent past such that the respondents’ top gift card choice company does not appear last in the ranking (that is, is not the company that advertises least on misinformation websites) and in most cases, advertises more intensely on misinformation websites than its potential substitute in the same company category (for example, fast food, food delivery or ride-sharing). Such a treatment allows us to measure potential differences in the direction of consumers switching their gift card choices, such as switching towards companies that advertise more or less intensely on misinformation websites. It could also give consumers reasonable deniability such as “everyone advertises on misinformation websites” leading to ambiguous predictions about the exact impact of the treatment effect.

    Outcomes

    We measure two pre-registered behavioural outcomes that collectively allow us to measure how people respond to our information treatments in terms of both voice and exit25. After the information treatment, all participants are asked to make their final gift card choice from the same six options they were shown earlier. Our main outcome of interest is whether participants ‘exit’ or switch their gift card preference—that is, whether they select a different gift card after the information treatment than their top choice indicated before the information treatment. To ensure incentive compatibility, participants are (truthfully) told that those randomly selected to receive a gift card will be offered the gift card of their choice at the end of our study. As mentioned above, the probability of being randomly chosen to receive a gift card is 20%. We choose a high probability of receiving a gift card relative to other online experiments since prior work has shown that consumers process choice-relevant information more carefully as realization probability increases72. To make the gift card outcome as realistic as possible, we also had a large value gift card (US$25). The focus of our experiments is on single-shot outcomes. While it would have been interesting to capture longer-term effects, the cost of implementing our gift card outcome for a large sample and expenditure on the other studies made a follow-up study cost-prohibitive.

    Secondly, participants are given the option to sign one of several real online petitions that we made and hosted on Change.org. Participants can opt to sign a petition that advocates for either blocking or allowing advertising on misinformation or choose not to sign any petition. Further, participants could choose between two petitions for blocking advertisements on misinformation websites, suggesting that either: (1) advertising companies, or (2) digital advertising platforms, need to block advertisements from appearing on misinformation websites. Overall, participants selected among the following five choices: (1) “Companies like X need to block their ads from appearing on misinformation websites.”, where X is their top choice gift card company; (2) “Companies like X need to allow their ads to appear on misinformation websites.”, where X is their top choice gift card company; (3) “Digital ad platforms used by companies need to block ads from appearing on misinformation websites.”; (4) “Digital ad platforms used by companies need to allow ads to appear on misinformation websites.”; and (5) I do not want to sign any petition. To track the number of petition signatures for each of these four petition options across our randomized groups, we provide separate petition links to participants in each randomized group. We record several petition-related outcomes. First, we measure participants’ intention to sign a petition based on the option they select in this question. Participants who pass our attention check and opt to sign a petition are later provided with a link to their petition of choice. This allows tracking whether participants click on the petition link provided. Participants can also self-report whether they signed the petition. Finally, for each randomized group, we can track the total number of actual petition signatures.

    Our petition outcomes serves two purposes. While our gift card outcome measures how people change their consumption behaviour in response to the information provided, people may also respond to our information treat ments in alternative ways—for example, by voicing their concerns or supplying information to the parties involved25,26. Given that the process of signing a petition is costly, participants’ responses to this outcome would constitute a meaningful measure similar to petition measures used in prior experimental work73,74. Second, since participants must choose between signing either company or platform petitions, this outcome allows us to measure whether or not, across our treatments, people hold advertising companies more responsible for financing misinformation than the digital advertising platforms that automatically place advertisements for companies.

    In addition to our behavioural outcomes, we also record participants’ stated preferences. To do so, we ask participants about their degree of agreement with several statements about misinformation on a seven-point scale ranging from ‘strongly agree’ to ‘strongly disagree’. These include whether they think: (1) companies have an important role in reducing the spread of misinformation through their advertising practices; and whether (2) digital platforms should give companies the option to avoid advertising on misinformation websites.

    Heterogeneous treatment effects

    We explore heterogeneity in consumer responses along four pre-registered dimensions. First, prior research recognizes differences in the salience of prosocial motivations across gender75, with women being more affected by social-impact messages than men76 and more critical consumers of new media content77. Given these findings, we could expect female participants to be more strongly affected by our information treatments.

    Responses to our information treatments may also differ by respondents’ political orientation. According to prior research, conservatives are especially likely to associate the mainstream media with the term ‘fake news’. These perceptions are generally linked to lower trust in media, voting for Trump, and higher belief in conspiracy theories78. Moreover, conservatives are more likely to consume misinformation2 and the supply of misinformation has been found to be higher on the ideological right than on the left79. Consequently, we might expect stronger treatment effects for left-wing respondents.

    Consumers who more frequently use a company’s products or services could be presumed to be more loyal towards the company or derive greater utility from its use, which could limit changes in their behaviour37. Alternatively, more frequent consumers may be more strongly affected by our information treatments as they may perceive their usage as supporting such company practices to a greater extent than less frequent consumers.

    Finally, we measure whether people’s responses differ by whether they consume misinformation themselves based on whether they reported using misinformation outlets in the initial question asking them to select which news outlets they used in the past 12 months.

    Tackling experimental validity concerns

    In our incentivized, online setting where we measure behavioural outcomes, we expect experimenter demand effects to be minimal as has been evidenced in the experimental literature80,81. We take several steps to mitigate potential experimenter demand effects, including implementing best practices recommended in prior work9. First, our experiment has a neutral framing throughout the survey since the recruitment of participants. While recruiting participants, we invite them to “take a survey about the news, technology and businesses” without making any specific references to misinformation or its effects. While introducing misinformation websites and how they are identified by independent non-partisan organizations, we include examples of misinformation websites across the political spectrum (including both right-wing and left-wing sites) and provide an illustrative example of misinformation by foreign actors. In drafting the survey instruments, the phrasing of the questions and choices available were as neutral as possible. For example, while introducing our online petitions, we presented participants with the option to sign real petitions that suggest both blocking and allowing advertising on misinformation sites. Indeed, we find that the vast majority of participants believe that the information provided in the survey was unbiased as shown in Supplementary Fig. 4. Only about 10% of participants chose one of the ‘biased’ or ‘very biased’ options when asked to rate the political bias of the survey information provided from a seven-point scale ranging from ‘very right-wing biased’ to ‘very left-wing biased’.

    In our active control design, participants in all randomized groups are presented with the same baseline information about misinformation, given misinformation-related information in the information intervention and asked the same questions after the information intervention to emphasize the same topics and minimize potential differences in the understanding of the study across treatment groups. Moreover, to maximize privacy and increase truthful reporting82, respondents complete the surveys on their own devices without the physical presence of a researcher. We also do not collect respondents’ names or contact details (with the exception of eliciting emails to provide gift cards to participants at the end of the study).

    In presenting our information interventions and measuring our behavioural outcomes, we take special care to not highlight the names of the specific entities being randomized across groups to avoid emphasizing what is being measured. We do, however, highlight our gift card incentives by putting the gift card information in bold text to ensure incentive compatibility since prior work has found that failing to make incentives conspicuous can vastly undermine their ability to shift behaviour83.

    Apart from making the above design choices to minimize experimenter demand effects, we measure their relevance using a survey question. Since demand effects are less likely a concern if participants cannot identify the intent of the study9, we ask participants an open-ended question—that is, “What do you think is the purpose of our study?”. Following prior work84,85, we then analyse the responses to this question to examine whether they differ across treatment groups. To measure potential differences in the respondents’ perceptions of the study, we examine their open-ended text responses about the purpose of the study using a Support Vector Machine classifier, which incorporates several features in text analysis, including word, character, and sentence counts, sentiments, topics (using Gensim) and word embeddings. We predict treatment status using the classifier, keeping 75% of the sample for the training set and the remaining 25% as the test set. The classifier predicts treatment status similar to chance for our main treatment groups relative to the control group, as shown in Supplementary Table 11. These results, which are similar in magnitude to those found in previous research84,85, suggest that our treatments do not substantially affect participants’ perceptions about the purpose of the study. Overall, this analysis gives us confidence that our main experimental findings are unlikely to be driven by experimenter demand effects.

    To address external validity concerns, we incorporate additional exit outcomes in the paper, showing that treated individuals switched to lower preference products (Table 1, columns 3 and 4) and products across categories (Table 1, columns 5 and 6) after our information interventions by 8 and 5 percentage points, respectively. We also show in Supplementary Table 8 that as the difference between participants’ highest weighted and second highest weighted gift card choice increases, their switching behaviour decreases. This shows that the weights assigned by participants to their gift card options are capturing meaningful and costly differences in value, highlighting the external validity of our findings. More generally, our pre-registered heterogeneity analysis lends credence to the study’s external validity. In line with expectations, we find that less frequent users and more politically liberal individuals are likelier to switch (see Extended Data Table 3 for the full set of pre-registered heterogeneity results). Moreover, we find that the cost of switching gift cards varies based on participants’ observable characteristics. For example, treated participants who reported not using any of the misinformation news outlets in our survey lost 50% of the median value (US$12.50) of their initial top choice gift card whereas treated participants who reported reading such outlets lost 33.3% of the median value (US$8.33) of their initial top choice gift card. Participants’ text responses also indicate that they believed their choices to be consequential (see Supplementary Tables 1 and 2). As an example, while explaining their choice of gift card, one participant stated, “Because I would most likely use this gift card on my next visit to… and it is less likely that i would use the others.” Regarding the petition outcome, one participant stated “The source of this problem seems to be from the digital advertising platforms, so I’d rather sign the petition that stops them from putting ads on misinformation websites.”

    Decision-maker experiment design

    We followed the same IRB review, pre-registration and consent procedures as those used for our consumer study. This study addresses two research questions. First, we aim to measure the existing beliefs and preferences decision-makers have about advertising on misinformation websites. This will help inform whether companies may be inadvertently or willingly sustaining online misinformation. Secondly, we ask: how do decision-makers update their beliefs and demand for a platform-based solution to avoid advertising on misinformation websites in response to information about the role of platforms in amplifying the financing of misinformation? This will suggest whether companies may be more interested in adopting advertising platforms that reduce the financing of misinformation. To this end, we conduct an information-provision experiment9. While past work has examined how firm behaviour regarding market decisions changes in response to new information48,49, it is unclear how information on the role of digital advertising platforms in amplifying advertising on misinformation would affect decision-makers’ non-market strategies.

    Setting and sample recruitment

    To recruit participants, we partnered with the executive education programmes at the Stanford Graduate School of Business and Heinz College at Carnegie Mellon University. We did so in order to survey senior managers and leaders who could influence strategic decision-making within their firms, in contrast to studies relying heavily on MBA students for understanding decision-making in various contexts such as competition, pricing, strategic alliances and marketing86,87,88,89. Additionally, partnering with two university programmes instead of a specific firm allowed us to access a more diverse sample of companies than prior work that sampled specific types of firms—for example, innovative firms, startups or small businesses90,91,92. Throughout this study, we use the preferences of decision-makers (for example, chief executive officers) as a proxy for company-level preferences since people in such roles shape the outcomes of their companies through their strategic decisions93,94.

    Our partner organizations sent emails to their alumni on our behalf. We used neutral language in our study recruitment emails to attract a broad audience of participants to our survey regardless of their initial beliefs and concerns about misinformation, stating our goal as “conducting vital research on the role of digital technologies in impacting your organization” without mentioning misinformation. We received 567 complete responses, of which 90% are kept since they are from currently employed respondents. To ensure data quality, we dropped an additional 13% of responses where participants were inattentive in answering the survey, resulting in a final sample of 442 responses. These participants were determined to be inattentive since they provided an answer greater than 100 when asked to estimate a number out of 100 in the two questions eliciting their prior beliefs about companies and platforms before the information treatment was provided. Our final sample of 442 respondents is from companies that span all the 23 industries in our descriptive analysis. Moreover, as shown in Supplementary Fig. 5, our sample of participants represents a broad array of company sizes and experience levels at their current roles. Additionally, about 22% of the executives in our sample (and 25% of all our participants) are women, which is aligned with the 21% to 26% industry estimates of women in senior roles globally95,96.

    Supplementary Fig. 2 shows the design of the survey study. We first elicit participants’ current employment status. All those working in some capacity are allowed to continue the survey, whereas the rest of the participants are screened out. After asking for their main occupation, all participants in the experiment are provided with baseline information on misinformation and advertising similar to that provided in the consumer experiment.

    Baseline beliefs and preferences

     In our pre-registration, we highlighted that we would measure the baseline beliefs and preferences of decision-makers. We measure participants’ baseline beliefs about the roles of companies in general, their own company and platforms in general in financing misinformation. Specifically, participants are asked to estimate the number of companies among the most active 100 advertisers whose advertisements appeared on misinformation websites during the past three years (2019–2021). Additionally, we ask participants to report whether they think their company or organization had its advertisements appear on misinformation websites in the past three years. Finally, we measure participants’ beliefs about the role of digital advertising platforms in placing advertisements on misinformation websites. To do so, we first inform participants that during the past three years (2019–2021), out of every 100 companies that did not use digital advertising platforms, eight companies appeared on misinformation websites on average. We then asked participants to provide their best estimate for the number of companies whose advertisements appeared on misinformation websites out of every 100 companies that did use digital advertising platforms.

    In addition to recording participants’ stated preferences using self-reported survey measures, we measure participants’ revealed preferences. To ensure incentive compatibility, participants are asked three questions in a randomized order: (1) information demand about consumer responses—that is, whether they would like to learn how consumers respond to companies whose advertisements appear on misinformation websites (based on our consumer survey experiment); (2) advertisement check—that is, whether they would like to know about their own company’s advertisements appearing on misinformation websites in the recent past; and (3) demand for a solution—that is, whether they would like to sign up for a 15-minute information session on how companies can manage where their advertisements appear online. Participants are told they can receive information about consumer responses at the end of the study if they opt to receive it whereas the advertisement check and solution information are provided as a follow-up after the survey. Participants are required to provide their emails and company name for the advertisement check. To sign up for an information session from our industry partner on a potential solution to avoid advertising on misinformation websites, participants sign up on a separate form by providing their emails. Since all three types of information offered are novel and otherwise costly to obtain, we expect respondents’ demand for such information to capture their revealed preferences.

    Information intervention

    Participants are then randomized into a treatment group, which receives information about the role of digital advertising platforms in placing advertising on misinformation websites, and a control group, which does not receive this information. Based on the dataset we assembled, participants are given factual information that companies that used digital advertising platforms were about ten times more likely to appear on misinformation websites than companies that did not use such platforms in the recent past. This information is identical to the information provided to participants in the T2 (that is, platform only) group in the consumer experiment.

    Outcomes

    After the information intervention, we first measure participants’ posterior beliefs about the role of digital advertising platforms in placing advertisements on misinformation websites following our pre-registration. Participants are told about the average number of companies whose advertisements appear per month on misinformation websites that are not monetized by digital advertising platforms. They are then asked to estimate the average number of companies whose advertisements appear monthly on misinformation websites that use digital advertising platforms. This question measures whether participants believe that the use of digital advertising platforms amplifies advertising on misinformation websites.

    We record two behavioural outcomes, which were pre-registered as our primary outcomes of interest after the information intervention. Our main outcome of interest is the respondents’ demand for a platform-based solution to avoid advertising on misinformation websites. Participants can opt to learn more about two different types of information—that is: (1) which platforms least frequently place companies’ advertising on misinformation websites; and (2) which types of analytics technologies are used to improve advertising performance—or opt not to receive any information. Since participants can only opt to receive one of the two types of information, this question is meant to capture the trade-off between respondents’ concern for avoiding misinformation outlets and their desire to improve advertising performance, respectively. Participants are told that they will be provided with the information they choose at the end of this study. Following the literature in measuring information acquisition97, we measure respondents’ demand for solution information, which serves as a revealed-preference proxy for their interest in implementing a solution for their organization.

    Additionally, to measure whether the information treatment increases concern for financing misinformation in general, we record a second behavioural measure. Participants are told that the research team will donate US$100 to one of two organizations after randomly selecting one of the first hundred responses: (1) the GDI; and (2) DataKind, which helps mission-driven organizations increase their impact by unlocking their data science potential ethically and responsibly.

    Tackling experimental validity concerns

    Similarly to our consumer experiment, this survey was carried out in an online setting, where experimenter demand effects are limited80,81. We followed best practices9 by keeping the treatment language neutral and ensuring the anonymity of the participants wherever possible. We find that most participants believe that the information provided in the survey was unbiased. Only about 7% of participants chose one of the ‘biased’ or ‘very biased’ options when asked to rate the political bias of the survey information provided from a seven-point scale ranging from ‘very right-wing biased’ to ‘very left-wing biased’.

    Importantly, to ensure truthful reporting, our main experimental outcomes were incentive-compatible. In particular, respondents who chose our platform solution demand outcome to learn about which platforms least contribute to placing companies’ advertisements on misinformation websites had to face a trade-off between receiving this information and receiving information on improving advertising performance. Additionally, our baseline information demand outcomes elicited before the information intervention were also incentive-compatible in that participants would be asked to follow up on their decisions whether they opted for additional information via email or via an online information session.

    These design choices are made to minimize demand effects on our main outcomes of interest. However, it is possible that these effects are still relevant, partially because participants may have an interest in ‘doing the right thing’ on a survey administered by an institution they have a connection with. We measure the relevance of potential demand effects using a survey question mirroring the approach used for our consumer experiment. To measure potential differences in the respondents’ perceptions of the study across our treatment and control groups, we predict treatment status based on respondents’ open-ended text responses about the purpose of the study via a support vector machine classifier, keeping 75% of the sample for the training set and the remaining 25% as the test set. We find that the classifier is only slightly worse than random chance in predicting treatment status (Supplementary Table 16) but similar in magnitude to those in the consumer experiment. Therefore, although experimenter demand effects may still be present, these results suggest that these effects do not drive our findings.

    We address the external validity of our findings by verifying the decision-making capacity of our respondents within their organizations and by examining the generalizability of our sample. We find that the vast majority of those whose job titles we verify (94%) serve in executive or managerial roles within their organizations. The regression estimates in Supplementary Tables 18 and 19 show that our results remain qualitatively and quantitatively similar after the exclusion of the small sample of individuals in non-executive and non-managerial roles. Moreover, the verified and self-reported decision-makers are similar across observable characteristics as reported in Supplementary Table 17, suggesting limited selection in our verification process. To examine the generalizability of our sample, we investigate their observable characteristics. As shown in Supplementary Fig. 5, our sample of participants represents a broad array of company sizes and experience levels at their current roles. Additionally, about 22% of the executives in our sample (and 25% of all our participants) are women, which is aligned with the 21% to 26% industry estimates of women in senior roles globally95,96.

    Reporting summary

    Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.

    [ad_2]

    Source link

  • The economic commitment of climate change

    [ad_1]

    Historical climate data

    Historical daily 2-m temperature and precipitation totals (in mm) are obtained for the period 1979–2019 from the W5E5 database. The W5E5 dataset comes from ERA-5, a state-of-the-art reanalysis of historical observations, but has been bias-adjusted by applying version 2.0 of the WATCH Forcing Data to ERA-5 reanalysis data and precipitation data from version 2.3 of the Global Precipitation Climatology Project to better reflect ground-based measurements49,50,51. We obtain these data on a 0.5° × 0.5° grid from the Inter-Sectoral Impact Model Intercomparison Project (ISIMIP) database. Notably, these historical data have been used to bias-adjust future climate projections from CMIP-6 (see the following section), ensuring consistency between the distribution of historical daily weather on which our empirical models were estimated and the climate projections used to estimate future damages. These data are publicly available from the ISIMIP database. See refs. 7,8 for robustness tests of the empirical models to the choice of climate data reanalysis products.

    Future climate data

    Daily 2-m temperature and precipitation totals (in mm) are taken from 21 climate models participating in CMIP-6 under a high (RCP8.5) and a low (RCP2.6) greenhouse gas emission scenario from 2015 to 2100. The data have been bias-adjusted and statistically downscaled to a common half-degree grid to reflect the historical distribution of daily temperature and precipitation of the W5E5 dataset using the trend-preserving method developed by the ISIMIP50,52. As such, the climate model data reproduce observed climatological patterns exceptionally well (Supplementary Table 5). Gridded data are publicly available from the ISIMIP database.

    Historical economic data

    Historical economic data come from the DOSE database of sub-national economic output53. We use a recent revision to the DOSE dataset that provides data across 83 countries, 1,660 sub-national regions with varying temporal coverage from 1960 to 2019. Sub-national units constitute the first administrative division below national, for example, states for the USA and provinces for China. Data come from measures of gross regional product per capita (GRPpc) or income per capita in local currencies, reflecting the values reported in national statistical agencies, yearbooks and, in some cases, academic literature. We follow previous literature3,7,8,54 and assess real sub-national output per capita by first converting values from local currencies to US dollars to account for diverging national inflationary tendencies and then account for US inflation using a US deflator. Alternatively, one might first account for national inflation and then convert between currencies. Supplementary Fig. 12 demonstrates that our conclusions are consistent when accounting for price changes in the reversed order, although the magnitude of estimated damages varies. See the documentation of the DOSE dataset for further discussion of these choices. Conversions between currencies are conducted using exchange rates from the FRED database of the Federal Reserve Bank of St. Louis55 and the national deflators from the World Bank56.

    Future socio-economic data

    Baseline gridded gross domestic product (GDP) and population data for the period 2015–2100 are taken from the middle-of-the-road scenario SSP2 (ref. 15). Population data have been downscaled to a half-degree grid by the ISIMIP following the methodologies of refs. 57,58, which we then aggregate to the sub-national level of our economic data using the spatial aggregation procedure described below. Because current methodologies for downscaling the GDP of the SSPs use downscaled population to do so, per-capita estimates of GDP with a realistic distribution at the sub-national level are not readily available for the SSPs. We therefore use national-level GDP per capita (GDPpc) projections for all sub-national regions of a given country, assuming homogeneity within countries in terms of baseline GDPpc. Here we use projections that have been updated to account for the impact of the COVID-19 pandemic on the trajectory of future income, while remaining consistent with the long-term development of the SSPs59. The choice of baseline SSP alters the magnitude of projected climate damages in monetary terms, but when assessed in terms of percentage change from the baseline, the choice of socio-economic scenario is inconsequential. Gridded SSP population data and national-level GDPpc data are publicly available from the ISIMIP database. Sub-national estimates as used in this study are available in the code and data replication files.

    Climate variables

    Following recent literature3,7,8, we calculate an array of climate variables for which substantial impacts on macroeconomic output have been identified empirically, supported by further evidence at the micro level for plausible underlying mechanisms. See refs. 7,8 for an extensive motivation for the use of these particular climate variables and for detailed empirical tests on the nature and robustness of their effects on economic output. To summarize, these studies have found evidence for independent impacts on economic growth rates from annual average temperature, daily temperature variability, total annual precipitation, the annual number of wet days and extreme daily rainfall. Assessments of daily temperature variability were motivated by evidence of impacts on agricultural output and human health, as well as macroeconomic literature on the impacts of volatility on growth when manifest in different dimensions, such as government spending, exchange rates and even output itself7. Assessments of precipitation impacts were motivated by evidence of impacts on agricultural productivity, metropolitan labour outcomes and conflict, as well as damages caused by flash flooding8. See Extended Data Table 1 for detailed references to empirical studies of these physical mechanisms. Marked impacts of daily temperature variability, total annual precipitation, the number of wet days and extreme daily rainfall on macroeconomic output were identified robustly across different climate datasets, spatial aggregation schemes, specifications of regional time trends and error-clustering approaches. They were also found to be robust to the consideration of temperature extremes7,8. Furthermore, these climate variables were identified as having independent effects on economic output7,8, which we further explain here using Monte Carlo simulations to demonstrate the robustness of the results to concerns of imperfect multicollinearity between climate variables (Supplementary Methods Section 2), as well as by using information criteria (Supplementary Table 1) to demonstrate that including several lagged climate variables provides a preferable trade-off between optimally describing the data and limiting the possibility of overfitting.

    We calculate these variables from the distribution of daily, d, temperature, Tx,d, and precipitation, Px,d, at the grid-cell, x, level for both the historical and future climate data. As well as annual mean temperature, \({\bar{T}}_{x,y}\), and annual total precipitation, Px,y, we calculate annual, y, measures of daily temperature variability, \({\widetilde{T}}_{x,y}\):

    $${\widetilde{T}}_{x,y}=\frac{1}{12}\mathop{\sum }\limits_{m=1}^{12}\sqrt{\frac{1}{{D}_{m}}\mathop{\sum }\limits_{d=1}^{{D}_{m}}{({T}_{x,d,m,y}-{\bar{T}}_{x,m})}^{2}},$$

    (1)

    the number of wet days, Pwdx,y:

    $${{\rm{Pwd}}}_{x,y}=\mathop{\sum }\limits_{d=1}^{{D}_{y}}H\left({P}_{x,d}-1\,{\rm{mm}}\right)$$

    (2)

    and extreme daily rainfall:

    $${{\rm{Pext}}}_{x,y}=\mathop{\sum }\limits_{d=1}^{{D}_{y}}H\left({P}_{x,d}-{P99.9}_{x}\right)\times {P}_{x,d},$$

    (3)

    in which Tx,d,m,y is the grid-cell-specific daily temperature in month m and year y, \({\bar{T}}_{x,m,{y}}\) is the year and grid-cell-specific monthly, m, mean temperature, Dm and Dy the number of days in a given month m or year y, respectively, H the Heaviside step function, 1 mm the threshold used to define wet days and P99.9x is the 99.9th percentile of historical (1979–2019) daily precipitation at the grid-cell level. Units of the climate measures are degrees Celsius for annual mean temperature and daily temperature variability, millimetres for total annual precipitation and extreme daily precipitation, and simply the number of days for the annual number of wet days.

    We also calculated weighted standard deviations of monthly rainfall totals as also used in ref. 8 but do not include them in our projections as we find that, when accounting for delayed effects, their effect becomes statistically indistinct and is better captured by changes in total annual rainfall.

    Spatial aggregation

    We aggregate grid-cell-level historical and future climate measures, as well as grid-cell-level future GDPpc and population, to the level of the first administrative unit below national level of the GADM database, using an area-weighting algorithm that estimates the portion of each grid cell falling within an administrative boundary. We use this as our baseline specification following previous findings that the effect of area or population weighting at the sub-national level is negligible7,8.

    Empirical model specification: fixed-effects distributed lag models

    Following a wide range of climate econometric literature16,60, we use panel regression models with a selection of fixed effects and time trends to isolate plausibly exogenous variation with which to maximize confidence in a causal interpretation of the effects of climate on economic growth rates. The use of region fixed effects, μr, accounts for unobserved time-invariant differences between regions, such as prevailing climatic norms and growth rates owing to historical and geopolitical factors. The use of yearly fixed effects, ηy, accounts for regionally invariant annual shocks to the global climate or economy such as the El Niño–Southern Oscillation or global recessions. In our baseline specification, we also include region-specific linear time trends, kry, to exclude the possibility of spurious correlations resulting from common slow-moving trends in climate and growth.

    The persistence of climate impacts on economic growth rates is a key determinant of the long-term magnitude of damages. Methods for inferring the extent of persistence in impacts on growth rates have typically used lagged climate variables to evaluate the presence of delayed effects or catch-up dynamics2,18. For example, consider starting from a model in which a climate condition, Cr,y, (for example, annual mean temperature) affects the growth rate, Δlgrpr,y (the first difference of the logarithm of gross regional product) of region r in year y:

    $${\Delta {\rm{lgrp}}}_{r,y}={\mu }_{r}+{\eta }_{y}+{k}_{r}y+\alpha {C}_{r,y}+{\varepsilon }_{r,y},$$

    (4)

    which we refer to as a ‘pure growth effects’ model in the main text. Typically, further lags are included,

    $${\Delta {\rm{lgrp}}}_{r,y}={\mu }_{r}+{\eta }_{y}+{k}_{r}y+\mathop{\sum }\limits_{L=0}^{{\rm{NL}}}{\alpha }_{L}{C}_{r,y-L}+{\varepsilon }_{r,y},$$

    (5)

    and the cumulative effect of all lagged terms is evaluated to assess the extent to which climate impacts on growth rates persist. Following ref. 18, in the case that,

    $$\mathop{\sum }\limits_{L=0}^{{\rm{NL}}}{\alpha }_{L} < 0\,{\rm{for}}\,{\alpha }_{0} < 0\,{\rm{or}}\,\mathop{\sum }\limits_{L=0}^{{\rm{NL}}}{\alpha }_{L} > 0\,{\rm{for}}\,{\alpha }_{0} > 0,$$

    (6)

    the implication is that impacts on the growth rate persist up to NL years after the initial shock (possibly to a weaker or a stronger extent), whereas if

    $$\mathop{\sum }\limits_{L=0}^{{\rm{NL}}}{\alpha }_{L}=0,$$

    (7)

    then the initial impact on the growth rate is recovered after NL years and the effect is only one on the level of output. However, we note that such approaches are limited by the fact that, when including an insufficient number of lags to detect a recovery of the growth rates, one may find equation (6) to be satisfied and incorrectly assume that a change in climatic conditions affects the growth rate indefinitely. In practice, given a limited record of historical data, including too few lags to confidently conclude in an infinitely persistent impact on the growth rate is likely, particularly over the long timescales over which future climate damages are often projected2,24. To avoid this issue, we instead begin our analysis with a model for which the level of output, lgrpr,y, depends on the level of a climate variable, Cr,y:

    $${{\rm{lgrp}}}_{r,y}={\mu }_{r}+{\eta }_{y}+{k}_{r}y+\alpha {C}_{r,y}+{\varepsilon }_{r,y}.$$

    (8)

    Given the non-stationarity of the level of output, we follow the literature19 and estimate such an equation in first-differenced form as,

    $${\Delta {\rm{lgrp}}}_{r,y}={\mu }_{r}+{\eta }_{y}+{k}_{r}y+\alpha {\Delta C}_{r,y}+{\varepsilon }_{r,y},$$

    (8)

    which we refer to as a model of ‘pure level effects’ in the main text. This model constitutes a baseline specification in which a permanent change in the climate variable produces an instantaneous impact on the growth rate and a permanent effect only on the level of output. By including lagged variables in this specification,

    $${\Delta {\rm{lgrp}}}_{r,y}={\mu }_{r}+{\eta }_{y}+{k}_{r}y+\mathop{\sum }\limits_{L=0}^{{\rm{NL}}}{\alpha }_{L}{\Delta C}_{r,y-L}+{\varepsilon }_{r,y},$$

    (9)

    we are able to test whether the impacts on the growth rate persist any further than instantaneously by evaluating whether αL > 0 are statistically significantly different from zero. Even though this framework is also limited by the possibility of including too few lags, the choice of a baseline model specification in which impacts on the growth rate do not persist means that, in the case of including too few lags, the framework reverts to the baseline specification of level effects. As such, this framework is conservative with respect to the persistence of impacts and the magnitude of future damages. It naturally avoids assumptions of infinite persistence and we are able to interpret any persistence that we identify with equation (9) as a lower bound on the extent of climate impact persistence on growth rates. See the main text for further discussion of this specification choice, in particular about its conservative nature compared with previous literature estimates, such as refs. 2,18.

    We allow the response to climatic changes to vary across regions, using interactions of the climate variables with historical average (1979–2019) climatic conditions reflecting heterogenous effects identified in previous work7,8. Following this previous work, the moderating variables of these interaction terms constitute the historical average of either the variable itself or of the seasonal temperature difference, \({\hat{T}}_{r}\), or annual mean temperature, \({\bar{T}}_{r}\), in the case of daily temperature variability7 and extreme daily rainfall, respectively8.

    The resulting regression equation with N and M lagged variables, respectively, reads:

    $$\begin{array}{l}{\Delta {\rm{lgrp}}}_{r,y}\,=\,{\mu }_{r}+{\eta }_{y}+{k}_{r}y+\mathop{\sum }\limits_{L=0}^{N}({{\alpha }_{1,L}\Delta \bar{T}}_{r,y-L}+{\alpha }_{2,L}{\Delta \bar{T}}_{r,y-L}\times {\bar{T}}_{r})\\ \,\,\,\,\,+\mathop{\sum }\limits_{L=0}^{N}({{\alpha }_{3,L}\Delta \widetilde{T}}_{r,y-L}+{\alpha }_{4,L}{\Delta \widetilde{T}}_{r,y-L}\times {\widehat{T}}_{r})\\ \,\,\,\,\,+\mathop{\sum }\limits_{L=0}^{M}({\alpha }_{5,L}\Delta {P}_{r,y-L}+{\alpha }_{6,L}\Delta {P}_{r,y-L}\times {P}_{r})\\ \,\,\,\,\,+\mathop{\sum }\limits_{L=0}^{M}({\alpha }_{7,L}\Delta {{\rm{Pwd}}}_{r,y-L}+{\alpha }_{8,L}\Delta {{\rm{Pwd}}}_{r,y-L}\times {{\rm{Pwd}}}_{r})\\ \,\,\,\,\,+\mathop{\sum }\limits_{L=0}^{M}({\alpha }_{9,L}\Delta {{\rm{Pext}}}_{r,y-L}+{\alpha }_{10,L}\Delta {{\rm{Pext}}}_{r,y-L}\times {\bar{T}}_{r})+{{\epsilon }}_{r,y}\end{array}$$

    (10)

    in which Δlgrpr,y is the annual, regional GRPpc growth rate, measured as the first difference of the logarithm of real GRPpc, following previous work2,3,7,8,18,19. Fixed-effects regressions were run using the fixest package in R (ref. 61).

    Estimates of the coefficients of interest αi,L are shown in Extended Data Fig. 1 for N = M = 10 lags and for our preferred choice of the number of lags in Supplementary Figs. 1–3. In Extended Data Fig. 1, errors are shown clustered at the regional level, but for the construction of damage projections, we block-bootstrap the regressions by region 1,000 times to provide a range of parameter estimates with which to sample the projection uncertainty (following refs. 2,31).

    Spatial-lag model

    In Supplementary Fig. 14, we present the results from a spatial-lag model that explores the potential for climate impacts to ‘spill over’ into spatially neighbouring regions. We measure the distance between centroids of each pair of sub-national regions and construct spatial lags that take the average of the first-differenced climate variables and their interaction terms over neighbouring regions that are at distances of 0–500, 500–1,000, 1,000–1,500 and 1,500–2000 km (spatial lags, ‘SL’, 1 to 4). For simplicity, we then assess a spatial-lag model without temporal lags to assess spatial spillovers of contemporaneous climate impacts. This model takes the form:

    $$\begin{array}{c}{\Delta {\rm{l}}{\rm{g}}{\rm{r}}{\rm{p}}}_{r,y}\,=\,{\mu }_{r}+{\eta }_{y}+{k}_{r}y+\mathop{\sum }\limits_{{\rm{S}}{\rm{L}}=0}^{N}({\alpha }_{1,{\rm{S}}{\rm{L}}}\Delta {\bar{T}}_{r-{\rm{S}}{\rm{L}},y}+{\alpha }_{2,{\rm{S}}{\rm{L}}}\Delta {\bar{T}}_{r-{\rm{S}}{\rm{L}},y}\times {\bar{T}}_{r-{\rm{S}}{\rm{L}}})\\ \,\,\,\,\,+\mathop{\sum }\limits_{{\rm{S}}{\rm{L}}=0}^{N}({\alpha }_{3,{\rm{S}}{\rm{L}}}\Delta {\mathop{T}\limits^{ \sim }}_{r-{\rm{S}}{\rm{L}},y}+{\alpha }_{4,{\rm{S}}{\rm{L}}}\Delta {\mathop{T}\limits^{ \sim }}_{r-{\rm{S}}{\rm{L}},y}\times {\hat{T}}_{r-{\rm{S}}{\rm{L}}})\\ \,\,\,\,\,+\mathop{\sum }\limits_{{\rm{S}}{\rm{L}}=0}^{N}({\alpha }_{5,{\rm{S}}{\rm{L}}}\Delta {P}_{r-{\rm{S}}{\rm{L}},y}+{\alpha }_{6,{\rm{S}}{\rm{L}}}\Delta {P}_{r-{\rm{S}}{\rm{L}},y}\times {P}_{r-{\rm{S}}{\rm{L}}})\\ \,\,\,\,\,+\mathop{\sum }\limits_{{\rm{S}}{\rm{L}}=0}^{N}({\alpha }_{7,{\rm{S}}{\rm{L}}}\Delta {{\rm{P}}{\rm{w}}{\rm{d}}}_{r-{\rm{S}}{\rm{L}},y}+{\alpha }_{8,{\rm{S}}{\rm{L}}}\Delta {{\rm{P}}{\rm{w}}{\rm{d}}}_{r-{\rm{S}}{\rm{L}},y}\times {{\rm{P}}{\rm{w}}{\rm{d}}}_{r-{\rm{S}}{\rm{L}}})\\ \,\,\,\,\,+\mathop{\sum }\limits_{{\rm{S}}{\rm{L}}=0}^{N}({\alpha }_{9,{\rm{S}}{\rm{L}}}\Delta {{\rm{P}}{\rm{e}}{\rm{x}}{\rm{t}}}_{r-{\rm{S}}{\rm{L}},y}+{\alpha }_{10,{\rm{S}}{\rm{L}}}\Delta {{\rm{P}}{\rm{e}}{\rm{x}}{\rm{t}}}_{r-{\rm{S}}{\rm{L}},y}\times {\bar{T}}_{r-{\rm{S}}{\rm{L}}})+{{\epsilon }}_{r,y},\end{array}$$

    (11)

    in which SL indicates the spatial lag of each climate variable and interaction term. In Supplementary Fig. 14, we plot the cumulative marginal effect of each climate variable at different baseline climate conditions by summing the coefficients for each climate variable and interaction term, for example, for average temperature impacts as:

    $${\rm{M}}{\rm{E}}=\mathop{\sum }\limits_{{\rm{S}}{\rm{L}}=0}^{N}({\alpha }_{1,{\rm{S}}{\rm{L}}}+{\alpha }_{2,{\rm{S}}{\rm{L}}}{\bar{T}}_{r-{\rm{S}}{\rm{L}}}).$$

    (12)

    These cumulative marginal effects can be regarded as the overall spatially dependent impact to an individual region given a one-unit shock to a climate variable in that region and all neighbouring regions at a given value of the moderating variable of the interaction term.

    Constructing projections of economic damage from future climate change

    We construct projections of future climate damages by applying the coefficients estimated in equation (10) and shown in Supplementary Tables 2–4 (when including only lags with statistically significant effects in specifications that limit overfitting; see Supplementary Methods Section 1) to projections of future climate change from the CMIP-6 models. Year-on-year changes in each primary climate variable of interest are calculated to reflect the year-to-year variations used in the empirical models. 30-year moving averages of the moderating variables of the interaction terms are calculated to reflect the long-term average of climatic conditions that were used for the moderating variables in the empirical models. By using moving averages in the projections, we account for the changing vulnerability to climate shocks based on the evolving long-term conditions (Supplementary Figs. 10 and 11 show that the results are robust to the precise choice of the window of this moving average). Although these climate variables are not differenced, the fact that the bias-adjusted climate models reproduce observed climatological patterns across regions for these moderating variables very accurately (Supplementary Table 6) with limited spread across models (<3%) precludes the possibility that any considerable bias or uncertainty is introduced by this methodological choice. However, we impose caps on these moderating variables at the 95th percentile at which they were observed in the historical data to prevent extrapolation of the marginal effects outside the range in which the regressions were estimated. This is a conservative choice that limits the magnitude of our damage projections.

    Time series of primary climate variables and moderating climate variables are then combined with estimates of the empirical model parameters to evaluate the regression coefficients in equation (10), producing a time series of annual GRPpc growth-rate reductions for a given emission scenario, climate model and set of empirical model parameters. The resulting time series of growth-rate impacts reflects those occurring owing to future climate change. By contrast, a future scenario with no climate change would be one in which climate variables do not change (other than with random year-to-year fluctuations) and hence the time-averaged evaluation of equation (10) would be zero. Our approach therefore implicitly compares the future climate-change scenario to this no-climate-change baseline scenario.

    The time series of growth-rate impacts owing to future climate change in region r and year y, δr,y, are then added to the future baseline growth rates, πr,y (in log-diff form), obtained from the SSP2 scenario to yield trajectories of damaged GRPpc growth rates, ρr,y. These trajectories are aggregated over time to estimate the future trajectory of GRPpc with future climate impacts:

    $${{\rm{GRPpc}}}_{r,Y}={{\rm{GRPpc}}}_{r,2020}\mathop{\sum }\limits_{y=2020}^{Y}{\rho }_{r,y}={{\rm{GRPpc}}}_{r,2020}\mathop{\sum }\limits_{y=2020}^{Y}\left(1+{\pi }_{r,y}+{\delta }_{r,y}\right),$$

    (13)

    in which GRPpcr,y=2020 is the initial log level of GRPpc. We begin damage estimates in 2020 to reflect the damages occurring since the end of the period for which we estimate the empirical models (1979–2019) and to match the timing of mitigation-cost estimates from most IAMs (see below).

    For each emission scenario, this procedure is repeated 1,000 times while randomly sampling from the selection of climate models, the selection of empirical models with different numbers of lags (shown in Supplementary Figs. 1–3 and Supplementary Tables 2–4) and bootstrapped estimates of the regression parameters. The result is an ensemble of future GRPpc trajectories that reflect uncertainty from both physical climate change and the structural and sampling uncertainty of the empirical models.

    Estimates of mitigation costs

    We obtain IPCC estimates of the aggregate costs of emission mitigation from the AR6 Scenario Explorer and Database hosted by IIASA23. Specifically, we search the AR6 Scenarios Database World v1.1 for IAMs that provided estimates of global GDP and population under both a SSP2 baseline and a SSP2-RCP2.6 scenario to maintain consistency with the socio-economic and emission scenarios of the climate damage projections. We find five IAMs that provide data for these scenarios, namely, MESSAGE-GLOBIOM 1.0, REMIND-MAgPIE 1.5, AIM/GCE 2.0, GCAM 4.2 and WITCH-GLOBIOM 3.1. Of these five IAMs, we use the results only from the first three that passed the IPCC vetting procedure for reproducing historical emission and climate trajectories. We then estimate global mitigation costs as the percentage difference in global per capita GDP between the SSP2 baseline and the SSP2-RCP2.6 emission scenario. In the case of one of these IAMs, estimates of mitigation costs begin in 2020, whereas in the case of two others, mitigation costs begin in 2010. The mitigation cost estimates before 2020 in these two IAMs are mostly negligible, and our choice to begin comparison with damage estimates in 2020 is conservative with respect to the relative weight of climate damages compared with mitigation costs for these two IAMs.

    [ad_2]

    Source link

  • Global supply chains amplify economic costs of future extreme heat risk

    [ad_1]

    Our methodology, in essence, combines three modules of climate, health and economy with full validation (Extended Data Fig. 1). The integrated model links climate module (estimating future climate parameters including surface air temperature and relative humidity and so on), demographic and health module (simulating future world population dynamics and exposure–response functions to warming) and economic module (dynamic footprint of heat-induced labour loss on global economy and supply chain).

    Climate module

    Fourteen GCMs involved in the framework of CMIP6 (Extended Data Table 1) with ten bias-corrected models from ISIMIP3b49,50 are used to estimate the modelled heat stress projection for the end of the twenty-first century. Five models were randomly averaged several times from the climate model ensemble as a Monte Carlo uncertainty analysis. ERA5 re-analysis data51 from 1985 to 2022 are used for bias-correction and validation. Climatic parameters such as maximum and average temperature and relative humidity on a daily scale are integrated, which are closely related to future working environment (Supplementary Fig. 9).

    Many institutes, including International Standards Organization (ISO) and US National Institute for Occupational Safety and Health (NIOSH), use WBGT to quantify different amounts of heat stress and define the percentage of a typical working hour that a person can work while maintaining core body temperature. To facilitate the long-term calculation, we use18 simplified WBGT, which approximates WBGT well using temperature (Ta) and relative humidity (RH)52,53 as parameters such as solar radiation and wind speed have higher uncertainty and weaker effects at the global scale. To take into account indoor heat exposures for industrial and service sector workers, we used the approximation that indoor WGBTindoor = WBGToutdoor − 4, based on a deduction of the radiation exposure factor from the formula below18:

    $${{\rm{WBGT}}}_{{\rm{outdoor}}}=0.567\times {T}_{{\rm{a}}}+3.94+0.393\times E$$

    (1)

    $$E=\frac{{\rm{RH}}}{100}\times 6.105\times \exp \left(17.27\times \frac{{T}_{{\rm{a}}}}{\left(237.7+{T}_{{\rm{a}}}\right)}\right)$$

    (2)

    We also calculated the spatial and temporal evolutionary trends in the occurrence of future heatwaves to calculate excess mortality. There is no consistent definition for heatwave worldwide because people may have acclimatized to their local climatic zones and different studies have applied various temperature metrics54,55. Heatwaves are usually defined by absolute or relative temperature threshold in consecutive days56. There are various ways to define a heatwave. For example, the IPCC defines heatwave as “a period of abnormally hot weather, often defined with reference to a relative temperature threshold, lasting from two days to months”, whereas the Chinese Meteorological Administration defined heatwave as “at least three consecutive days with maximum temperature exceeding 35 °C”. Others31 identified heatwave using the TX90p criterion, that is, when the 90th percentile of the distribution of regional maximum temperatures spanned by data from the period 1981–2010 was exceeded for at least three consecutive days. In our study, two or more consecutive days above the 95% threshold of the 1985–2015 ERA5 daily mean temperature51,57 were defined as a heatwave, which is considered to be a moderate estimation and is widely used in epidemiological studies36,58,59. Several definitions, such as four or more consecutive days above the 97.5% threshold, are used as sensitivity analysis. Considering certain amounts of climate adaptation of the local resident along the warming climate, dynamic heatwave thresholds60 are defined as part of the uncertainty analysis in this study; that is two or more consecutive days above the 95% threshold of the daily mean temperature between 1985 and the year before the target year were defined as a heatwave (ERA5 data are used for 1985–2014; climate projection data are used after 2015). The use of a dynamic threshold based on both historical and climate projections data helps to incorporate the human adaptation of heat stress in a long-term warming scenario, as reported in recent studies61,62,63,64.

    Health costs related to heat exposure

    Some studies have shown that the health impact of heatwaves could vary substantially with location65,66. Few studies have investigated the heatwave-induced mortality risk at a global scale41,67. A primitive health risk function associating heatwave mortality risks with four different climate zones was established by ref. 36 on the basis of a comprehensive study using data from 400 communities in 18 countries/regions across several years (1972–2012). Here, we used the relative risk coefficients (Extended Data Table 2) from figure 4 of ref. 36 for four different climate zones (Extended Data Fig. 3) to estimate potential heatwave-related death due to climate change on a global scale. The simplified four-climate-zone-based estimation may neglect subregional characters and should be interpreted with caution, as further factors affecting heat-induced death (such as air condition accessibility68, age69,70,71,72 and humidity73) are not included in this study.

    The number of excess deaths Dhw during a heatwave period was calculated at each grid cell level (0.5°) with the following equation:

    $${D}_{{\rm{hw}}}={\rm{POP}}\times {\rm{MR}}\times \left({\rm{RR}}-1\right)\times {\rm{HWN}}$$

    (3)

    POP is the population at the given location consistent with the SSPs74. MR is the average daily mortality rate (2009–2019) at the country level obtained from the World Bank75. For 37 countries with large territory and more refined data (for example, European Union (including UK), Russia, Ukraine, China, the USA, Canada, Brazil, South Africa, India and Australia), we used state/provincial statistics based on data from national statistical offices (Source, World Bank; state/province level data for European Union, Eurostat76; Russia, The Russian Fertility and Mortality database77; China, China Statistical Yearbook 201978; the USA, National Institutes of Health79; Brazil, Fundação Amazônia de Amparo a Estudos e Pesquisas80; Canada, Statistics Canada81; Australia, Australian Bureau of Statistics82; India, Ministry of Finance Economic Survey83). RR is the relative risk of mortality caused by heatwaves. HWN is the number of heatwave days for the given year and location (Extended Data Fig. 2).

    The calculated excess deaths are translated to a social-economic loss on the basis of the value of statistical life (VSL). The concept of VSL is widely used throughout the world to monetize fatality risks in benefit–cost analyses. The VSL represents the individual’s local money–mortality risk tradeoff value, which is the value of small changes in risk, not the value attached to identified lives. The country-based VSL estimation used in this research is adopted from the global health risks pricing study by ref. 84. The estimation is based on the estimated VSL in the USA (US$201911 million) and coupled with an income elasticity of 1.0 to adjust the VSL to other countries using the fixed-effects specification. A similar health valuation method has been adopted in past studies85,86 and was recommended in the report of the World Bank87. Moreover, a sensitivity test is conducted under the assumption that all life would be valued equally across the world (Supplementary Figs. 2 and 3). For such a test, an averaged VSL is calculated by summing up each country’s income-based VSL times its population then dividing by the total population of the world.

    Expose function of labour productivity

    The increase in daily temperatures affects the efficiency of workers and reduces safe working time. A compromise in endurance capacity due to thermoregulatory stress was already evident at 21 °C. Different studies used similar methods to evaluate the labour loss function. The form of logistic function with ‘S’ shape has become the consensus of the academic community but the specific functional equation and parameters are various in different studies. The loss functions used in mainstream research include exponential function88 as equation (4), cumulative normal distribution function5,41 as equation (5) and so on. In this research, we adopt the cumulative normal distribution function (equation (5)) as our benchmark function because it was extensively applied and case proven in 3-year reports of the Lancet Countdown on health and climate change5,41,89,90. Because the Hothaps function (equation (4)) is subject to parameter uncertainty as a result of being based on a few empirical studies, we use it to test for the sensitivity of our estimates (Supplementary Figs. 4 and 5). Our methodology identifies three ISO standard work intensity amounts: 200 W (assumed to be office workers in the service industry, engaged in light work indoors), 300 W (assumed to be industrial workers, engaged in moderate work indoors) and 400 W (assumed to be construction or agricultural workers, engaged in heavy work outside). For example, to calculate workability loss fraction in India’s food production sector (300 W, indoor), we bring the corresponding parameters (Extended Data Table 3) and WBGTindoor into equation (5). Previous studies have tended to ignore indoor workforce loss, assuming that the indoor workforce was very low under current climate condition or protected by air conditioning91. However, a growing number of studies have proved that future indoor labour losses cannot be underestimated31. For example, only 7% of households in India possess an air conditioner, despite having extremely high cooling needs. Considering the severe adaptation cooling deficit in emerging economies92, indoor labour losses must be fully considered in global-scale studies. This study uses the climate–income–air conditioner usage function published by ref. 93 to assess the rate of air conditioning protection in conjunction with the per capita income of each country under each SSP scenario. Higher per capita income in each country leads to higher air-conditioning penetration, whereas the climate base determines the rate and trend of increase in air-conditioning penetration (elasticity of penetration to income). In our study, we improved the function by replacing cooling degree days (CDDs) with indoor WBGT, as CDDs only consider temperature neglecting humidity. Only the indoor workforce under air conditioning, will be protected from heat-induced loss.

    $${\text{Workability}}_{\text{Hothaps}}=0.1+\frac{0.9}{\left(1+{\left(\frac{{\rm{WBGT}}}{{\alpha }_{1}}\right)}^{{\alpha }_{2}}\right)}$$

    (4)

    $$\text{Loss fraction}\,=\,\frac{1}{2}\left(1+{\rm{E}}{\rm{R}}{\rm{F}}(\frac{\text{WBGT}-{\text{Prod}}_{\text{mean}}}{{{\rm{P}}{\rm{r}}{\rm{o}}{\rm{d}}}_{{\rm{S}}{\rm{D}}}\times \sqrt{2}})\right)$$

    (5)

    Of which the parameters for a given activity level (Prodmean and ProdSD, defined as the amount of internal heat generated in performing the activity) are given in Extended Data Table 3, and ERF is the error function defined as:

    $${\rm{ERF}}\left(z\right)=\frac{2}{\sqrt{{\rm{\pi }}}}{\int }_{0}^{z}{e}^{-{t}^{2}}{\rm{d}}t$$

    (6)

    To calculate average daily impacts, we use an approximation for hourly data based on the 4 + 4 + 4 method implemented by ref. 14. We assume that 4 h per day is close to WBGTmax and 4 h per day is close to WBGTmean (early morning and early evening). The remaining 4 h of a 12 h daylight day is assumed to be halfway between WBGTmean and WBGTmax (labelled WBGThalf). The analysis above gives the summer daily potential workability lost in each grid cell at each amount of work intensity and environment (200–400 W, indoor or outdoor). By combining this with the dynamic population grid under each SSP scenario (see Supplementary Fig. 13 for comparison with static population setting), we aggregate to obtain country-scale labour productivity losses. In the disaster footprint model, we adopt the approach presented by ref. 5 which defines the timeframe for computing labour productivity losses as the warm season (June to 30 September in the Northern Hemisphere and December to 30 March in the Southern Hemisphere) to adjust the overestimation of the risk of moderate hot temperature, as the model is more applicable to sudden and strong shocks rather than moderate changes throughout the year.

    Global disaster footprint analysis module

    The global economic loss will be calculated using the following hybrid input–output and computable general equilibrium (CGE) global trade module. Our global trade module is an extension of the adaptive regional input–output (ARIO) model20,94,95, which was widely used in the literature to simulate the propagation of negative shocks throughout the economy96,97,98,99. Our model improves the ARIO model in two ways. The first improvement is related to the substitutability of products from the same sector sourced from different regions. Second, in our model, clients will choose their suppliers across regions on the basis of their capacity. These two improvements contribute to a more realistic representation of bottlenecks along global supply chains100.

    Our global trade module mainly includes four modules: production module, allocation module, demand module and simulation module. The production module is mainly designed for characterizing the firm’s production activities. The allocation module is mainly used to describe how firms allocate output to their clients, including downstream firms (intermediate demand) and households (final demand). The demand module is mainly used to describe how clients place orders to their suppliers. And the simulation module is mainly designed for executing the whole simulation procedure.

    Production module

    The production module is used to characterize production processes. Firms rent capital and use labour to process natural resources and intermediate inputs produced by other firms into a specific product. The production process for firm i can be expressed as follows,

    $${x}_{i}=f({\rm{for}}\,{\rm{all}}\,p,{z}_{i}^{{\rm{p}}}\,;{{\rm{va}}}_{i})$$

    where xi denotes the output of the firm i, in monetary value; p denotes type of intermediate products; \({z}_{i}^{{\rm{p}}}\) denotes intermediate products used in production processes; vai denotes the primary inputs to production, such as labour (L), capital (K) and natural resources (NR). The production function for firms is f(·). There is a wide range of functional forms, such as Leontief101, Cobb–Douglas and constant elasticity of substitution production function102. Different functional forms reflect the possibility for firms to substitute an input for another. Considering that heat stress tends to be concentrated in a specific short period of time, during which economic agents cannot easily replace inputs as suitable substitutes, might temporarily be unavailable, we use Leontief production function which does not allow substitution between inputs.

    $${x}_{i}=\min \left({\rm{for}}\;{\rm{all}}\,p,\frac{{z}_{i}^{{\rm{p}}}}{{a}_{i}^{{\rm{p}}}}\,;\frac{{{\rm{va}}}_{i}}{{b}_{i}}\right)$$

    where \({a}_{i}^{{\rm{p}}}\) and \({b}_{i}\) are the input coefficients calculated as

    $${a}_{i}^{{\rm{p}}}=\frac{{\bar{z}}_{i}^{{\rm{p}}}}{{\bar{x}}_{i}}$$

    and

    $${b}_{i}=\frac{{\overline{{\rm{va}}}}_{i}}{{\bar{x}}_{i}}$$

    where the horizontal bar indicates the value of that variable in the equilibrium state. In an equilibrium state, producers use intermediate products and primary inputs to produce goods and services to satisfy demand from their clients. After a disaster, output will decline. From a production perspective, there are mainly the following constraints.

    Labour supply constraints

    Labour constraints during heat stress or after a disaster may impose severe knock-on effects on the rest of the economy21,103. This makes labour constraints a key factor to consider in disaster impact analysis. For example, in the case of heat stress, these constraints can arise from employees’ inability to work as a result of illness or extreme environmental temperatures beyond health threshold. In this model, the proportion of surviving productive capacity from the constrained labour productive capacity (\({x}_{i}^{{\rm{L}}}\)) after a shock is defined as:

    $${x}_{i}^{{\rm{L}}}(t)=(1-{\gamma }_{i}^{{\rm{L}}}(t))\times {\bar{x}}_{i}$$

    Where \({\gamma }_{i}^{{\rm{L}}}(t)\) is the proportion of labour that is unavailable at each time step t during heat stress; \((1-{\gamma }_{i}^{{\rm{L}}}(t))\) contains the available proportion of employment at time t.

    $${\gamma }_{i}^{{\rm{L}}}(t)=\left({\bar{L}}_{i}-{L}_{i}(t)\right)/{\bar{L}}_{i}$$

    The proportion of the available productive capacity of labour is thus a function of the losses from the sectoral labour forces and its predisaster employment level. Following the assumption of the fixed proportion of production functions, the productive capacity of labour in each region after a disaster (\({x}_{i}^{{\rm{L}}}\)) will represent a linear proportion of the available labour capacity at each time step. Take heatwaves as an example; during extreme heatwaves that last for days on end, governments and businesses often shut down work to reduce the risk of serious illnesses such as pyrexia. This imposes an exogenous negative shock on the economic network.

    Constraints on productive capital

    Similar to labour constraints, the productive capacity of industrial capital in each region during the aftermath of a disaster (\({x}_{i}^{{\rm{K}}}\)) will be constrained by the surviving capacity of the industrial capital30,96,104,105,106. The share of damage to each sector is directly considered as the proportion of the monetized damage to capital assets in relation to the total value of industrial capital for each sector, which is disclosed in the event account vector for each region \(({\gamma }_{i}^{{\rm{K}}})\), following ref. 107. This assumption is embodied in the essence of the input–output model, which is hard-coded through the Leontief-type production function and its restricted substitution. As capital and labour are considered perfectly complementary as well as the main production factors and the full employment of those factors in the economy is also assumed, we assume that damage in capital assets is directly related with production level and, therefore, VA level. Then, the remaining productive capacity of the industrial capital at each time step is defined as:

    $${x}_{i}^{{\rm{K}}}(t)=(1-{\gamma }_{i}^{{\rm{K}}}(t))\times {\bar{x}}_{i}$$

    Where, \({\bar{K}}_{i}\) is the capital stock of firm \(i\) in the predisaster situation and Ki(t) is the surviving capital stock of firm \(i\) at time \(t\) during the recovery process

    $${\gamma }_{i}^{{\rm{K}}}(t)=\left({\bar{K}}_{i}-{K}_{i}(t)\right)/{\bar{K}}_{i}$$

    Supply constraints

    Firms will purchase intermediate products from their supplier in each period. Insufficient inventory of a firm’s intermediate products will create a bottleneck for production activities. The potential production level that the inventory of the pth intermediate product can support is

    $${x}_{i}^{{\rm{p}}}(t)=\frac{{S}_{i}^{{\rm{p}}}(t-1)}{{a}_{i}^{{\rm{p}}}}$$

    where \({S}_{i}^{{\rm{p}}}(t-1)\) refers to the amount of pth intermediate products held by firm i at the end of time step t − 1.

    Considering all the limitation mentioned above, the maximum supply capacity of firm i can be expressed as

    $${x}_{i}^{\max }\left(t\right)=\min \left({x}_{i}^{{\rm{L}}}\left(t\right)\,;{x}_{i}^{{\rm{K}}}\left(t\right)\,;\,{\rm{for}}\;{\rm{all}}\,p,{x}_{i}^{{\rm{p}}}\left(t\right)\right)$$

    The actual production of firm i, \({x}_{i}^{{\rm{a}}}(t)\), depends on both its maximum supply capacity and the total orders the firm received from its clients, \({{\rm{TD}}}_{i}(t-1)\) (see section on the ‘Demand module’),

    $${x}_{i}^{{\rm{a}}}\left(t\right)=\min \left({x}_{i}^{\max }\left(t\right),{{\rm{TD}}}_{i}(t-1)\right)$$

    The inventory held by firm i will be consumed during the production process,

    $${S}_{i}^{{\rm{p}},{\rm{u}}{\rm{s}}{\rm{e}}{\rm{d}}}(t)={a}_{i}^{p}\times {x}_{i}^{{\rm{a}}}(t)$$

    Allocation module

    The allocation module mainly describes how suppliers allocate products to their clients. When some firms in the economic system suffer a negative shock, their production will be constrained by a shortage to primary inputs such as a shortage of labour supply during extreme heat stress. In this case, a firm’s output will not be able to fill all orders of its clients. A rationing scheme that reflects a mechanism on the basis of which a firm allocates an insufficient amount of products to its clients is needed108. For this case study, we applied a proportional rationing scheme according to which a firm allocates its output in proportion to its orders. Under the proportional rationing scheme, the amounts of products of firm i allocated to firm j, \({{\rm{F}}{\rm{R}}{\rm{C}}}_{j}^{i}\) and household h, \({{\rm{H}}{\rm{R}}{\rm{C}}}_{h}^{i}\) are as follows,

    $${{\rm{F}}{\rm{R}}{\rm{C}}}_{j}^{i}(t)=\frac{{{\rm{F}}{\rm{O}}{\rm{D}}}_{i}^{j}(t-1)}{({\sum }_{j}{{\rm{F}}{\rm{O}}{\rm{D}}}_{i}^{j}(t-1)+{\sum }_{h}{{\rm{H}}{\rm{O}}{\rm{D}}}_{i}^{h}(t-1))}\times {x}_{i}^{{\rm{a}}}(t)$$

    $${{\rm{H}}{\rm{R}}{\rm{C}}}_{h}^{i}(t)=\frac{{{\rm{H}}{\rm{O}}{\rm{D}}}_{i}^{h}(t-1)}{({\sum }_{j}{{\rm{F}}{\rm{O}}{\rm{D}}}_{i}^{j}(t-1)+{\sum }_{h}{{\rm{H}}{\rm{O}}{\rm{D}}}_{i}^{h}(t-1))}\times {x}_{i}^{{\rm{a}}}(t)$$

    where \({{\rm{F}}{\rm{O}}{\rm{D}}}_{i}^{j}(t-1)\) refers to the order issued by firm j to its supplier i in time step t − 1, and \({{\rm{H}}{\rm{O}}{\rm{D}}}_{i}^{h}(t-1)\) refers to the order issued by household h to its supplier j. Firm j received intermediates to restore its inventories,

    $${S}_{j}^{p,{\rm{restored}}}\left(t\right)={\sum }_{i\to p}{{\rm{FRC}}}_{j}^{i}(t)$$

    Therefore, the amount of intermediate p held by firm i at the end of period t is

    $${S}_{j}^{p}\left(t\right)={S}_{j}^{p}\left(t-1\right)-{S}_{j}^{p,{\rm{used}}}\left(t\right)+{S}_{j}^{p,{\rm{restored}}}$$

    Demand module

    The demand module represents a characterization of how firms and households issues orders to their suppliers at the end of each period. A firm orders its supplier because of the need to restore its intermediate product inventory. We assume that each firm has a specific target inventory level based on its maximum supply capacity in each time step,

    $${S}_{i}^{p,\ast }(t)={n}_{i}^{p}\times {{a}_{i}^{p}\times x}_{i}^{max}(t)$$

    Then the order issued by firm i to its supplier j is

    $${{\rm{F}}{\rm{O}}{\rm{D}}}_{j}^{i}(t)=\{\begin{array}{c}({S}_{i}^{p,\ast }(t)-{S}_{i}^{p}(t))\times \frac{{\bar{{\rm{F}}{\rm{O}}{\rm{D}}}}_{j}^{i}\times {x}_{j}^{a}(t)}{{\sum }_{j\to p}({\bar{{\rm{F}}{\rm{O}}{\rm{D}}}}_{j}^{i}\times {x}_{j}^{a}(t))},{\rm{i}}{\rm{f}}\,{S}_{i}^{p,\ast }(t) > {S}_{i}^{p}(t);\\ 0,\,{\rm{i}}{\rm{f}}\,{S}_{i}^{p,\ast }(t)\le {S}_{i}^{p}(t).\end{array}$$

    Households issue orders to their suppliers on the basis of their demand and the supply capacity of their suppliers. In this study, the demand of household h to final products q, \({{\rm{HD}}}_{h}^{q}\left(t\right)\), is given exogenously at each time step. Then, the order issued by household (HOD) h to its supplier j is

    $${{\rm{H}}{\rm{O}}{\rm{D}}}_{j}^{h}(t)={{\rm{H}}{\rm{D}}}_{h}^{q}(t)\times \frac{{\bar{{\rm{H}}{\rm{O}}{\rm{D}}}}_{j}^{h}\times {x}_{j}^{a}(t)}{{\sum }_{j\to q}({\bar{{\rm{H}}{\rm{O}}{\rm{D}}}}_{j}^{h}\times {x}_{j}^{a}(t))}$$

    The total order received (TOD) by firm j is

    $${{\rm{TOD}}}_{j}\left(t\right)={\sum }_{i}{{\rm{FOD}}}_{j}^{i}\left(t\right)+{\sum }_{h}{{\rm{HOD}}}_{j}^{h}\left(t\right)$$

    Simulation module

    At each time step, the actions of firms and households are as follows in Monte Carlo simulations.

    Firms plan and execute their production on the basis of three factors: (1) inventories of intermediate products they have, (2) supply of primary inputs and (3) orders from their clients. Firms will maximize their output under these constraints.

    Product allocation

    Firms allocate outputs to clients on the basis of their orders. In equilibrium, the output of firms just meets all orders. When production is constrained by exogenous negative shocks, outputs may not cover all orders. In this case, we use a proportional rationing scheme proposed in the literature20,108 (see section on ‘Allocation module’) to allocate products of firms.

    Firms and households issue orders to their suppliers for the next time step. Firms place orders with their suppliers on the basis of the gaps in their inventories (target inventory level minus existing inventory level). Households place orders with their suppliers on the basis of their demand. When a product comes from several suppliers, the allocation of orders is adjusted according to the production capacity of each supplier.

    This discrete-time dynamic procedure can reproduce the equilibrium of the economic system and can simulate the propagation of exogenous shocks, both from firm and household side or transportation disruptions, in the economic network. From the firm side, if the supply of a firm’s primary inputs is constrained, it will have two effects. On the one hand, the decline in output in this firm means that its clients’ orders cannot be fulfilled. This will result in a decrease in inventory of these clients, which will constrain their production. This is the so-called forward or downstream effect. On the other hand, less output in this firm also means less use of intermediate products from its suppliers. This will reduce the number of orders it places on its suppliers, which will further reduce the production level of its suppliers. This is the so-called backward or upstream effect. From the household side, the fluctuation of household demand caused by exogenous shocks will also trigger the aforementioned backward effect. Take tourism as an example, when the temperature is well beyond the comfort range of the visitor, the demand for tourism from households all over the world will decline significantly. This influence will further propagate to the accommodation and catering industry through supplier–client links.

    Economic footprint

    We define the VA decrease of all firms in a network caused by an exogenous negative shock as the disaster footprint of the shock. For the firm directly affected by exogenous negative shocks, its loss includes two parts: (1) the VA decrease caused by exogenous constraints and (2) the VA decrease caused by propagation. The former is the direct loss, whereas the latter is the indirect loss. A negative shock’s total economic footprint (TEFi,r), direct economic footprint (DEFi,r) and propagated economic footprint (PEFi,r) for firm i in region r are,

    $${{\rm{T}}{\rm{E}}{\rm{F}}}_{i,r}={\bar{{\rm{v}}{\rm{a}}}}_{i,r}\times T-{\sum }_{t=1}^{T}{{\rm{v}}{\rm{a}}}_{i,r}^{{\rm{a}}}(t)$$

    and,

    $${{\rm{D}}{\rm{E}}{\rm{F}}}_{i,r}={\bar{{\rm{v}}{\rm{a}}}}_{i,r}\times T-{\sum }_{t=1}^{T}{{\rm{v}}{\rm{a}}}_{i,r}^{max}(t)$$

    and,

    $${{\rm{PEF}}}_{i,r}={{\rm{TEF}}}_{i,r}-{{\rm{DEF}}}_{i,r}$$

    Global supply-chain network

    We build a global supply-chain network based on v.10 of the Global Trade Analysis Project (GTAP) database109 and use GTAP 9 (ref. 110), EMERGING database111 for robustness analysis. GTAP 10 provides a multiregional input–output (MRIO) table for the year 2014. Also, the database for the year 2011 was used for robustness testing. This MRIO table divides the world into 141 economies, each of which contains 65 production sectors (Supplementary Tables 4 and 5). If we treat each sector as a firm (producer) and assume that each region has a representative household, we can obtain the following information in the MRIO table: (1) suppliers and clients of each firm; (2) suppliers for each household and (3) the flow of each supplier–client connection under the equilibrium condition. This provides a benchmark for our model. We also used a dynamic CGE model consistent with the SSP scenarios for a parallel assessment and as part of the robustness check of the ARIO results. Specifically, the CGE model we used is a G-RDEM112 with aggregated ten regions and ten sectors113,114,115 (Supplementary Information section 1.3).

    When applying such a realistic and aggregated network to the disaster footprint model, we need to consider the substitutability of intermediate products supplied by suppliers from the same sector in different regions115,116,117. The substitution between some intermediate products is straightforward. For example, for a firm that extracts spices from bananas it does not make much of a difference if the bananas are sourced from the Philippines or Thailand. However, for a car manufacturing firm in Japan, which uses screws from Chinese auto parts suppliers and engines from German auto parts suppliers to assemble cars, the products of the suppliers in these two regions are non-substitutable. If we assume that all goods are non-substitutable as in the traditional input–output model, then we will overestimate the loss of producers such as the case of the fragrance extraction firm. If we assume that products from suppliers in the same sector can be completely substitutable, then we will substantially underestimate the losses of producers such as the Japanese car manufacturing firm. To alleviate these shortcomings in the evaluation of losses under the two assumptions, we allow for the possibility of substitution for each sector depending on the region and sector of the supplier (Supplementary Information section 1.3).

    Nonetheless, our estimates of economic damages from heat stress are subject to some important uncertainties118 and our methods may not capture all types of economic damages. We only include economic losses caused by heat stress on human activities without considering the impacts on infrastructure, crop growth and other factors. Considering the challenges of predicting changes to socioeconomic systems globally, we have followed the approach from the literature23,31,91,119 to simulate supply-chain indirect losses by considering the impact of future climate risks on current socioeconomic settings. We have not considered the potential substitution of labour with capital resulting from technological advances, such as mechanization. Our analysis ignores the different levels of trade openness and globalization among SSP narratives, as well as the role of dynamic factors such as technology and price. Again, although we have conducted robustness tests for different degrees of trade substitutability, the relevant parameter is set randomly in the Monte Carlo simulation rather than derived through a general equilibrium model. The results should therefore be interpreted with caution as indicating potential future climate change risks to the existing economy rather than as quantitative predictions, given that the static representation of the economic structure in our model inevitably skews the assessment in the long run.

    [ad_2]

    Source link

  • Artificial intelligence and illusions of understanding in scientific research

    [ad_1]

  • Crabtree, G. Self-driving laboratories coming of age. Joule 4, 2538–2541 (2020).

    Article 
    CAS 

    Google Scholar
     

  • Wang, H. et al. Scientific discovery in the age of artificial intelligence. Nature 620, 47–60 (2023). This review explores how AI can be incorporated across the research pipeline, drawing from a wide range of scientific disciplines.

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Dillion, D., Tandon, N., Gu, Y. & Gray, K. Can AI language models replace human participants? Trends Cogn. Sci. 27, 597–600 (2023).

    Article 
    PubMed 

    Google Scholar
     

  • Grossmann, I. et al. AI and the transformation of social science research. Science 380, 1108–1109 (2023). This forward-looking article proposes a variety of ways to incorporate generative AI into social-sciences research.

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Gil, Y. Will AI write scientific papers in the future? AI Mag. 42, 3–15 (2022).


    Google Scholar
     

  • Kitano, H. Nobel Turing Challenge: creating the engine for scientific discovery. npj Syst. Biol. Appl. 7, 29 (2021).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Benjamin, R. Race After Technology: Abolitionist Tools for the New Jim Code (Oxford Univ. Press, 2020). This book examines how social norms about race become embedded in technologies, even those that are focused on providing good societal outcomes.

  • Broussard, M. More Than a Glitch: Confronting Race, Gender, and Ability Bias in Tech (MIT Press, 2023).

  • Noble, S. U. Algorithms of Oppression: How Search Engines Reinforce Racism (New York Univ. Press, 2018).

  • Bender, E. M., Gebru, T., McMillan-Major, A. & Shmitchell, S. On the dangers of stochastic parrots: can language models be too big? in Proc. 2021 ACM Conference on Fairness, Accountability, and Transparency 610–623 (Association for Computing Machinery, 2021). One of the first comprehensive critiques of large language models, this article draws attention to a host of issues that ought to be considered before taking up such tools.

  • Crawford, K. Atlas of AI: Power, Politics, and the Planetary Costs of Artificial Intelligence (Yale Univ. Press, 2021).

  • Johnson, D. G. & Verdicchio, M. Reframing AI discourse. Minds Mach. 27, 575–590 (2017).

    Article 

    Google Scholar
     

  • Atanasoski, N. & Vora, K. Surrogate Humanity: Race, Robots, and the Politics of Technological Futures (Duke Univ. Press, 2019).

  • Mitchell, M. & Krakauer, D. C. The debate over understanding in AI’s large language models. Proc. Natl Acad. Sci. USA 120, e2215907120 (2023).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Kidd, C. & Birhane, A. How AI can distort human beliefs. Science 380, 1222–1223 (2023).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Birhane, A., Kasirzadeh, A., Leslie, D. & Wachter, S. Science in the age of large language models. Nat. Rev. Phys. 5, 277–280 (2023).

    Article 

    Google Scholar
     

  • Kapoor, S. & Narayanan, A. Leakage and the reproducibility crisis in machine-learning-based science. Patterns 4, 100804 (2023).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Hullman, J., Kapoor, S., Nanayakkara, P., Gelman, A. & Narayanan, A. The worst of both worlds: a comparative analysis of errors in learning from data in psychology and machine learning. In Proc. 2022 AAAI/ACM Conference on AI, Ethics, and Society (eds Conitzer, V. et al.) 335–348 (Association for Computing Machinery, 2022).

  • Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1, 206–215 (2019). This paper articulates the problems with attempting to explain AI systems that lack interpretability, and advocates for building interpretable models instead.

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Crockett, M. J., Bai, X., Kapoor, S., Messeri, L. & Narayanan, A. The limitations of machine learning models for predicting scientific replicability. Proc. Natl Acad. Sci. USA 120, e2307596120 (2023).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Lazar, S. & Nelson, A. AI safety on whose terms? Science 381, 138 (2023).

    Article 
    PubMed 

    Google Scholar
     

  • Collingridge, D. The Social Control of Technology (St Martin’s Press, 1980).

  • Wagner, G., Lukyanenko, R. & Paré, G. Artificial intelligence and the conduct of literature reviews. J. Inf. Technol. 37, 209–226 (2022).

    Article 

    Google Scholar
     

  • Hutson, M. Artificial-intelligence tools aim to tame the coronavirus literature. Nature https://doi.org/10.1038/d41586-020-01733-7 (2020).

    Article 
    PubMed 

    Google Scholar
     

  • Haas, Q. et al. Utilizing artificial intelligence to manage COVID-19 scientific evidence torrent with Risklick AI: a critical tool for pharmacology and therapy development. Pharmacology 106, 244–253 (2021).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Müller, H., Pachnanda, S., Pahl, F. & Rosenqvist, C. The application of artificial intelligence on different types of literature reviews – a comparative study. In 2022 International Conference on Applied Artificial Intelligence (ICAPAI) https://doi.org/10.1109/ICAPAI55158.2022.9801564 (Institute of Electrical and Electronics Engineers, 2022).

  • van Dinter, R., Tekinerdogan, B. & Catal, C. Automation of systematic literature reviews: a systematic literature review. Inf. Softw. Technol. 136, 106589 (2021).

    Article 

    Google Scholar
     

  • Aydın, Ö. & Karaarslan, E. OpenAI ChatGPT generated literature review: digital twin in healthcare. In Emerging Computer Technologies 2 (ed. Aydın, Ö.) 22–31 (İzmir Akademi Dernegi, 2022).

  • AlQuraishi, M. AlphaFold at CASP13. Bioinformatics 35, 4862–4865 (2019).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Jumper, J. et al. Highly accurate protein structure prediction with AlphaFold. Nature 596, 583–589 (2021).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Lee, J. S., Kim, J. & Kim, P. M. Score-based generative modeling for de novo protein design. Nat. Computat. Sci. 3, 382–392 (2023).

    Article 
    CAS 

    Google Scholar
     

  • Gómez-Bombarelli, R. et al. Design of efficient molecular organic light-emitting diodes by a high-throughput virtual screening and experimental approach. Nat. Mater. 15, 1120–1127 (2016).

    Article 
    PubMed 

    Google Scholar
     

  • Krenn, M. et al. On scientific understanding with artificial intelligence. Nat. Rev. Phys. 4, 761–769 (2022).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Extance, A. How AI technology can tame the scientific literature. Nature 561, 273–274 (2018).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Hastings, J. AI for Scientific Discovery (CRC Press, 2023). This book reviews current and future incorporation of AI into the scientific research pipeline.

  • Ahmed, A. et al. The future of academic publishing. Nat. Hum. Behav. 7, 1021–1026 (2023).

    Article 
    PubMed 

    Google Scholar
     

  • Gray, K., Yam, K. C., Zhen’An, A. E., Wilbanks, D. & Waytz, A. The psychology of robots and artificial intelligence. In The Handbook of Social Psychology (eds Gilbert, D. et al.) (in the press).

  • Argyle, L. P. et al. Out of one, many: using language models to simulate human samples. Polit. Anal. 31, 337–351 (2023).

    Article 

    Google Scholar
     

  • Aher, G., Arriaga, R. I. & Kalai, A. T. Using large language models to simulate multiple humans and replicate human subject studies. In Proc. 40th International Conference on Machine Learning (eds Krause, A. et al.) 337–371 (JMLR.org, 2023).

  • Binz, M. & Schulz, E. Using cognitive psychology to understand GPT-3. Proc. Natl Acad. Sci. USA 120, e2218523120 (2023).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Ornstein, J. T., Blasingame, E. N. & Truscott, J. S. How to train your stochastic parrot: large language models for political texts. Github, https://joeornstein.github.io/publications/ornstein-blasingame-truscott.pdf (2023).

  • He, S. et al. Learning to predict the cosmological structure formation. Proc. Natl Acad. Sci. USA 116, 13825–13832 (2019).

    Article 
    MathSciNet 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Mahmood, F. et al. Deep adversarial training for multi-organ nuclei segmentation in histopathology images. IEEE Trans. Med. Imaging 39, 3257–3267 (2020).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Teixeira, B. et al. Generating synthetic X-ray images of a person from the surface geometry. In Proc. IEEE Conference on Computer Vision and Pattern Recognition 9059–9067 (Institute of Electrical and Electronics Engineers, 2018).

  • Marouf, M. et al. Realistic in silico generation and augmentation of single-cell RNA-seq data using generative adversarial networks. Nat. Commun. 11, 166 (2020).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Watts, D. J. A twenty-first century science. Nature 445, 489 (2007).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • boyd, d. & Crawford, K. Critical questions for big data. Inf. Commun. Soc. 15, 662–679 (2012). This article assesses the ethical and epistemic implications of scientific and societal moves towards big data and provides a parallel case study for thinking about the risks of artificial intelligence.

    Article 

    Google Scholar
     

  • Jolly, E. & Chang, L. J. The Flatland fallacy: moving beyond low–dimensional thinking. Top. Cogn. Sci. 11, 433–454 (2019).

    Article 
    PubMed 

    Google Scholar
     

  • Yarkoni, T. & Westfall, J. Choosing prediction over explanation in psychology: lessons from machine learning. Perspect. Psychol. Sci. 12, 1100–1122 (2017).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Radivojac, P. et al. A large-scale evaluation of computational protein function prediction. Nat. Methods 10, 221–227 (2013).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Bileschi, M. L. et al. Using deep learning to annotate the protein universe. Nat. Biotechnol. 40, 932–937 (2022).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Barkas, N. et al. Joint analysis of heterogeneous single-cell RNA-seq dataset collections. Nat. Methods 16, 695–698 (2019).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Demszky, D. et al. Using large language models in psychology. Nat. Rev. Psychol. 2, 688–701 (2023).

    Article 

    Google Scholar
     

  • Karjus, A. Machine-assisted mixed methods: augmenting humanities and social sciences with artificial intelligence. Preprint at https://arxiv.org/abs/2309.14379 (2023).

  • Davies, A. et al. Advancing mathematics by guiding human intuition with AI. Nature 600, 70–74 (2021).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Peterson, J. C., Bourgin, D. D., Agrawal, M., Reichman, D. & Griffiths, T. L. Using large-scale experiments and machine learning to discover theories of human decision-making. Science 372, 1209–1214 (2021).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Ilyas, A. et al. Adversarial examples are not bugs, they are features. Preprint at https://doi.org/10.48550/arXiv.1905.02175 (2019)

  • Semel, B. M. Listening like a computer: attentional tensions and mechanized care in psychiatric digital phenotyping. Sci. Technol. Hum. Values 47, 266–290 (2022).

    Article 

    Google Scholar
     

  • Gil, Y. Thoughtful artificial intelligence: forging a new partnership for data science and scientific discovery. Data Sci. 1, 119–129 (2017).

    Article 

    Google Scholar
     

  • Checco, A., Bracciale, L., Loreti, P., Pinfield, S. & Bianchi, G. AI-assisted peer review. Humanit. Soc. Sci. Commun. 8, 25 (2021).

    Article 

    Google Scholar
     

  • Thelwall, M. Can the quality of published academic journal articles be assessed with machine learning? Quant. Sci. Stud. 3, 208–226 (2022).

    Article 

    Google Scholar
     

  • Dhar, P. Peer review of scholarly research gets an AI boost. IEEE Spectrum spectrum.ieee.org/peer-review-of-scholarly-research-gets-an-ai-boost (2020).

  • Heaven, D. AI peer reviewers unleashed to ease publishing grind. Nature 563, 609–610 (2018).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Conroy, G. How ChatGPT and other AI tools could disrupt scientific publishing. Nature 622, 234–236 (2023).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Nosek, B. A. et al. Replicability, robustness, and reproducibility in psychological science. Annu. Rev. Psychol. 73, 719–748 (2022).

    Article 
    PubMed 

    Google Scholar
     

  • Altmejd, A. et al. Predicting the replicability of social science lab experiments. PLoS ONE 14, e0225826 (2019).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Yang, Y., Youyou, W. & Uzzi, B. Estimating the deep replicability of scientific findings using human and artificial intelligence. Proc. Natl Acad. Sci. USA 117, 10762–10768 (2020).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Youyou, W., Yang, Y. & Uzzi, B. A discipline-wide investigation of the replicability of psychology papers over the past two decades. Proc. Natl Acad. Sci. USA 120, e2208863120 (2023).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Rabb, N., Fernbach, P. M. & Sloman, S. A. Individual representation in a community of knowledge. Trends Cogn. Sci. 23, 891–902 (2019). This comprehensive review paper documents the empirical evidence for distributed cognition in communities of knowledge and the resultant vulnerabilities to illusions of understanding.

    Article 
    PubMed 

    Google Scholar
     

  • Rozenblit, L. & Keil, F. The misunderstood limits of folk science: an illusion of explanatory depth. Cogn. Sci. 26, 521–562 (2002). This paper provided an empirical demonstration of the illusion of explanatory depth, and inspired a programme of research in cognitive science on communities of knowledge.

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Hutchins, E. Cognition in the Wild (MIT Press, 1995).

  • Lave, J. & Wenger, E. Situated Learning: Legitimate Peripheral Participation (Cambridge Univ. Press, 1991).

  • Kitcher, P. The division of cognitive labor. J. Philos. 87, 5–22 (1990).

    Article 

    Google Scholar
     

  • Hardwig, J. Epistemic dependence. J. Philos. 82, 335–349 (1985).

    Article 

    Google Scholar
     

  • Keil, F. in Oxford Studies In Epistemology (eds Gendler, T. S. & Hawthorne, J.) 143–166 (Oxford Academic, 2005).

  • Weisberg, M. & Muldoon, R. Epistemic landscapes and the division of cognitive labor. Philos. Sci. 76, 225–252 (2009).

    Article 

    Google Scholar
     

  • Sloman, S. A. & Rabb, N. Your understanding is my understanding: evidence for a community of knowledge. Psychol. Sci. 27, 1451–1460 (2016).

    Article 
    PubMed 

    Google Scholar
     

  • Wilson, R. A. & Keil, F. The shadows and shallows of explanation. Minds Mach. 8, 137–159 (1998).

    Article 

    Google Scholar
     

  • Keil, F. C., Stein, C., Webb, L., Billings, V. D. & Rozenblit, L. Discerning the division of cognitive labor: an emerging understanding of how knowledge is clustered in other minds. Cogn. Sci. 32, 259–300 (2008).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Sperber, D. et al. Epistemic vigilance. Mind Lang. 25, 359–393 (2010).

    Article 

    Google Scholar
     

  • Wilkenfeld, D. A., Plunkett, D. & Lombrozo, T. Depth and deference: when and why we attribute understanding. Philos. Stud. 173, 373–393 (2016).

    Article 

    Google Scholar
     

  • Sparrow, B., Liu, J. & Wegner, D. M. Google effects on memory: cognitive consequences of having information at our fingertips. Science 333, 776–778 (2011).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Fisher, M., Goddu, M. K. & Keil, F. C. Searching for explanations: how the internet inflates estimates of internal knowledge. J. Exp. Psychol. Gen. 144, 674–687 (2015).

    Article 
    PubMed 

    Google Scholar
     

  • De Freitas, J., Agarwal, S., Schmitt, B. & Haslam, N. Psychological factors underlying attitudes toward AI tools. Nat. Hum. Behav. 7, 1845–1854 (2023).

    Article 
    PubMed 

    Google Scholar
     

  • Castelo, N., Bos, M. W. & Lehmann, D. R. Task-dependent algorithm aversion. J. Mark. Res. 56, 809–825 (2019).

    Article 

    Google Scholar
     

  • Cadario, R., Longoni, C. & Morewedge, C. K. Understanding, explaining, and utilizing medical artificial intelligence. Nat. Hum. Behav. 5, 1636–1642 (2021).

    Article 
    PubMed 

    Google Scholar
     

  • Oktar, K. & Lombrozo, T. Deciding to be authentic: intuition is favored over deliberation when authenticity matters. Cognition 223, 105021 (2022).

    Article 
    PubMed 

    Google Scholar
     

  • Bigman, Y. E., Yam, K. C., Marciano, D., Reynolds, S. J. & Gray, K. Threat of racial and economic inequality increases preference for algorithm decision-making. Comput. Hum. Behav. 122, 106859 (2021).

    Article 

    Google Scholar
     

  • Claudy, M. C., Aquino, K. & Graso, M. Artificial intelligence can’t be charmed: the effects of impartiality on laypeople’s algorithmic preferences. Front. Psychol. 13, 898027 (2022).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Snyder, C., Keppler, S. & Leider, S. Algorithm reliance under pressure: the effect of customer load on service workers. Preprint at SSRN https://doi.org/10.2139/ssrn.4066823 (2022).

  • Bogert, E., Schecter, A. & Watson, R. T. Humans rely more on algorithms than social influence as a task becomes more difficult. Sci Rep. 11, 8028 (2021).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Raviv, A., Bar‐Tal, D., Raviv, A. & Abin, R. Measuring epistemic authority: studies of politicians and professors. Eur. J. Personal. 7, 119–138 (1993).

    Article 

    Google Scholar
     

  • Cummings, L. The “trust” heuristic: arguments from authority in public health. Health Commun. 29, 1043–1056 (2014).

    Article 
    PubMed 

    Google Scholar
     

  • Lee, M. K. Understanding perception of algorithmic decisions: fairness, trust, and emotion in response to algorithmic management. Big Data Soc. 5, https://doi.org/10.1177/2053951718756684 (2018).

  • Kissinger, H. A., Schmidt, E. & Huttenlocher, D. The Age of A.I. And Our Human Future (Little, Brown, 2021).

  • Lombrozo, T. Explanatory preferences shape learning and inference. Trends Cogn. Sci. 20, 748–759 (2016). This paper provides an overview of philosophical theories of explanatory virtues and reviews empirical evidence on the sorts of explanations people find satisfying.

    Article 
    PubMed 

    Google Scholar
     

  • Vrantsidis, T. H. & Lombrozo, T. Simplicity as a cue to probability: multiple roles for simplicity in evaluating explanations. Cogn. Sci. 46, e13169 (2022).

    Article 
    PubMed 

    Google Scholar
     

  • Johnson, S. G. B., Johnston, A. M., Toig, A. E. & Keil, F. C. Explanatory scope informs causal strength inferences. In Proc. 36th Annual Meeting of the Cognitive Science Society 2453–2458 (Cognitive Science Society, 2014).

  • Khemlani, S. S., Sussman, A. B. & Oppenheimer, D. M. Harry Potter and the sorcerer’s scope: latent scope biases in explanatory reasoning. Mem. Cognit. 39, 527–535 (2011).

    Article 
    PubMed 

    Google Scholar
     

  • Liquin, E. G. & Lombrozo, T. Motivated to learn: an account of explanatory satisfaction. Cogn. Psychol. 132, 101453 (2022).

    Article 
    PubMed 

    Google Scholar
     

  • Hopkins, E. J., Weisberg, D. S. & Taylor, J. C. V. The seductive allure is a reductive allure: people prefer scientific explanations that contain logically irrelevant reductive information. Cognition 155, 67–76 (2016).

    Article 
    PubMed 

    Google Scholar
     

  • Weisberg, D. S., Hopkins, E. J. & Taylor, J. C. V. People’s explanatory preferences for scientific phenomena. Cogn. Res. Princ. Implic. 3, 44 (2018).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Jerez-Fernandez, A., Angulo, A. N. & Oppenheimer, D. M. Show me the numbers: precision as a cue to others’ confidence. Psychol. Sci. 25, 633–635 (2014).

    Article 
    PubMed 

    Google Scholar
     

  • Kim, J., Giroux, M. & Lee, J. C. When do you trust AI? The effect of number presentation detail on consumer trust and acceptance of AI recommendations. Psychol. Mark. 38, 1140–1155 (2021).

    Article 

    Google Scholar
     

  • Nguyen, C. T. The seductions of clarity. R. Inst. Philos. Suppl. 89, 227–255 (2021). This article describes how reductive and quantitative explanations can generate a sense of understanding that is not necessarily correlated with actual understanding.

    Article 

    Google Scholar
     

  • Fisher, M., Smiley, A. H. & Grillo, T. L. H. Information without knowledge: the effects of internet search on learning. Memory 30, 375–387 (2022).

    Article 

    Google Scholar
     

  • Eliseev, E. D. & Marsh, E. J. Understanding why searching the internet inflates confidence in explanatory ability. Appl. Cogn. Psychol. 37, 711–720 (2023).

    Article 

    Google Scholar
     

  • Fisher, M. & Oppenheimer, D. M. Who knows what? Knowledge misattribution in the division of cognitive labor. J. Exp. Psychol. Appl. 27, 292–306 (2021).

    Article 
    PubMed 

    Google Scholar
     

  • Chromik, M., Eiband, M., Buchner, F., Krüger, A. & Butz, A. I think I get your point, AI! The illusion of explanatory depth in explainable AI. In 26th International Conference on Intelligent User Interfaces (eds Hammond, T. et al.) 307–317 (Association for Computing Machinery, 2021).

  • Strevens, M. No understanding without explanation. Stud. Hist. Philos. Sci. A 44, 510–515 (2013).

    Article 

    Google Scholar
     

  • Ylikoski, P. in Scientific Understanding: Philosophical Perspectives (eds De Regt, H. et al.) 100–119 (Univ. Pittsburgh Press, 2009).

  • Giudice, M. D. The prediction–explanation fallacy: a pervasive problem in scientific applications of machine learning. Preprint at PsyArXiv https://doi.org/10.31234/osf.io/4vq8f (2021).

  • Hofman, J. M. et al. Integrating explanation and prediction in computational social science. Nature 595, 181–188 (2021). This paper highlights the advantages and disadvantages of explanatory versus predictive approaches to modelling, with a focus on applications to computational social science.

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Shmueli, G. To explain or to predict? Stat. Sci. 25, 289–310 (2010).

    Article 
    MathSciNet 

    Google Scholar
     

  • Hofman, J. M., Sharma, A. & Watts, D. J. Prediction and explanation in social systems. Science 355, 486–488 (2017).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Logg, J. M., Minson, J. A. & Moore, D. A. Algorithm appreciation: people prefer algorithmic to human judgment. Organ. Behav. Hum. Decis. Process. 151, 90–103 (2019).

    Article 

    Google Scholar
     

  • Nguyen, C. T. Cognitive islands and runaway echo chambers: problems for epistemic dependence on experts. Synthese 197, 2803–2821 (2020).

    Article 

    Google Scholar
     

  • Breiman, L. Statistical modeling: the two cultures. Stat. Sci. 16, 199–215 (2001).

    Article 
    MathSciNet 

    Google Scholar
     

  • Gao, J. & Wang, D. Quantifying the benefit of artificial intelligence for scientific research. Preprint at arxiv.org/abs/2304.10578 (2023).

  • Hanson, B. et al. Garbage in, garbage out: mitigating risks and maximizing benefits of AI in research. Nature 623, 28–31 (2023).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Kleinberg, J. & Raghavan, M. Algorithmic monoculture and social welfare. Proc. Natl Acad. Sci. USA 118, e2018340118 (2021). This paper uses formal modelling methods to demonstrate that when companies all rely on the same algorithm to make decisions (an algorithmic monoculture), the overall quality of those decisions is reduced because valuable options can slip through the cracks, even when the algorithm performs accurately for individual companies.

    Article 
    MathSciNet 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Hofstra, B. et al. The diversity–innovation paradox in science. Proc. Natl Acad. Sci. USA 117, 9284–9291 (2020).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Hong, L. & Page, S. E. Groups of diverse problem solvers can outperform groups of high-ability problem solvers. Proc. Natl Acad. Sci. USA 101, 16385–16389 (2004).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Page, S. E. Where diversity comes from and why it matters? Eur. J. Soc. Psychol. 44, 267–279 (2014). This article reviews research demonstrating the benefits of cognitive diversity and diversity in methodological approaches for problem solving and innovation.

    Article 

    Google Scholar
     

  • Clarke, A. E. & Fujimura, J. H. (eds) The Right Tools for the Job: At Work in Twentieth-Century Life Sciences (Princeton Univ. Press, 2014).

  • Silva, V. J., Bonacelli, M. B. M. & Pacheco, C. A. Framing the effects of machine learning on science. AI Soc. https://doi.org/10.1007/s00146-022-01515-x (2022).

  • Sassenberg, K. & Ditrich, L. Research in social psychology changed between 2011 and 2016: larger sample sizes, more self-report measures, and more online studies. Adv. Methods Pract. Psychol. Sci. 2, 107–114 (2019).

    Article 

    Google Scholar
     

  • Simon, A. F. & Wilder, D. Methods and measures in social and personality psychology: a comparison of JPSP publications in 1982 and 2016. J. Soc. Psychol. https://doi.org/10.1080/00224545.2022.2135088 (2022).

    Article 
    PubMed 

    Google Scholar
     

  • Anderson, C. A. et al. The MTurkification of social and personality psychology. Pers. Soc. Psychol. Bull. 45, 842–850 (2019).

    Article 
    PubMed 

    Google Scholar
     

  • Latour, B. in The Social After Gabriel Tarde: Debates and Assessments (ed. Candea, M.) 145–162 (Routledge, 2010).

  • Porter, T. M. Trust in Numbers: The Pursuit of Objectivity in Science and Public Life (Princeton Univ. Press, 1996).

  • Lazer, D. et al. Meaningful measures of human society in the twenty-first century. Nature 595, 189–196 (2021).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Knox, D., Lucas, C. & Cho, W. K. T. Testing causal theories with learned proxies. Annu. Rev. Polit. Sci. 25, 419–441 (2022).

    Article 

    Google Scholar
     

  • Barberá, P. Birds of the same feather tweet together: Bayesian ideal point estimation using Twitter data. Polit. Anal. 23, 76–91 (2015).

    Article 

    Google Scholar
     

  • Brady, W. J., McLoughlin, K., Doan, T. N. & Crockett, M. J. How social learning amplifies moral outrage expression in online social networks. Sci. Adv. 7, eabe5641 (2021).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Barnes, J., Klinger, R. & im Walde, S. S. Assessing state-of-the-art sentiment models on state-of-the-art sentiment datasets. In Proc. 8th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis (eds Balahur, A. et al.) 2–12 (Association for Computational Linguistics, 2017).

  • Gitelman, L. (ed.) “Raw Data” is an Oxymoron (MIT Press, 2013).

  • Breznau, N. et al. Observing many researchers using the same data and hypothesis reveals a hidden universe of uncertainty. Proc. Natl Acad. Sci. USA 119, e2203150119 (2022). This study demonstrates how 73 research teams analysing the same dataset reached different conclusions about the relationship between immigration and public support for social policies, highlighting the subjectivity and uncertainty involved in analysing complex datasets.

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Gillespie, T. in Media Technologies: Essays on Communication, Materiality, and Society (eds Gillespie, T. et al.) 167–194 (MIT Press, 2014).

  • Leonelli, S. Data-Centric Biology: A Philosophical Study (Univ. Chicago Press, 2016).

  • Wang, A., Kapoor, S., Barocas, S. & Narayanan, A. Against predictive optimization: on the legitimacy of decision-making algorithms that optimize predictive accuracy. ACM J. Responsib. Comput., https://doi.org/10.1145/3636509 (2023).

  • Athey, S. Beyond prediction: using big data for policy problems. Science 355, 483–485 (2017).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • del Rosario Martínez-Ordaz, R. Scientific understanding through big data: from ignorance to insights to understanding. Possibility Stud. Soc. 1, 279–299 (2023).

    Article 

    Google Scholar
     

  • Nussberger, A.-M., Luo, L., Celis, L. E. & Crockett, M. J. Public attitudes value interpretability but prioritize accuracy in artificial intelligence. Nat. Commun. 13, 5821 (2022).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Zittrain, J. in The Cambridge Handbook of Responsible Artificial Intelligence: Interdisciplinary Perspectives (eds. Voeneky, S. et al.) 176–184 (Cambridge Univ. Press, 2022). This article articulates the epistemic risks of prioritizing predictive accuracy over explanatory understanding when AI tools are interacting in complex systems.

  • Shumailov, I. et al. The curse of recursion: training on generated data makes models forget. Preprint at arxiv.org/abs/2305.17493 (2023).

  • Latour, B. Science In Action: How to Follow Scientists and Engineers Through Society (Harvard Univ. Press, 1987). This book provides strategies and approaches for thinking about science as a social endeavour.

  • Franklin, S. Science as culture, cultures of science. Annu. Rev. Anthropol. 24, 163–184 (1995).

    Article 

    Google Scholar
     

  • Haraway, D. Situated knowledges: the science question in feminism and the privilege of partial perspective. Fem. Stud. 14, 575–599 (1988). This article acknowledges that the objective ‘view from nowhere’ is unobtainable: knowledge, it argues, is always situated.

    Article 

    Google Scholar
     

  • Harding, S. Objectivity and Diversity: Another Logic of Scientific Research (Univ. Chicago Press, 2015).

  • Longino, H. E. Science as Social Knowledge: Values and Objectivity in Scientific Inquiry (Princeton Univ. Press, 1990).

  • Daston, L. & Galison, P. Objectivity (Princeton Univ. Press, 2007). This book is a historical analysis of the shifting modes of ‘objectivity’ that scientists have pursued, arguing that objectivity is not a universal concept but that it shifts alongside scientific techniques and ambitions.

  • Prescod-Weinstein, C. Making Black women scientists under white empiricism: the racialization of epistemology in physics. Signs J. Women Cult. Soc. 45, 421–447 (2020).

    Article 

    Google Scholar
     

  • Mavhunga, C. What Do Science, Technology, and Innovation Mean From Africa? (MIT Press, 2017).

  • Schiebinger, L. The Mind Has No Sex? Women in the Origins of Modern Science (Harvard Univ. Press, 1991).

  • Martin, E. The egg and the sperm: how science has constructed a romance based on stereotypical male–female roles. Signs J. Women Cult. Soc. 16, 485–501 (1991). This case study shows how assumptions about gender affect scientific theories, sometimes delaying the articulation of what might be considered to be more accurate descriptions of scientific phenomena.

    Article 

    Google Scholar
     

  • Harding, S. Rethinking standpoint epistemology: What is “strong objectivity”? Centen. Rev. 36, 437–470 (1992). In this article, Harding outlines her position on ‘strong objectivity’, by which clearly articulating one’s standpoint can lead to more robust knowledge claims.


    Google Scholar
     

  • Oreskes, N. Why Trust Science? (Princeton Univ. Press, 2019). This book introduces the reader to 20 years of scholarship in science and technology studies, arguing that the tools the discipline has for understanding science can help to reinstate public trust in the institution.

  • Rolin, K., Koskinen, I., Kuorikoski, J. & Reijula, S. Social and cognitive diversity in science: introduction. Synthese 202, 36 (2023).

    Article 

    Google Scholar
     

  • Hong, L. & Page, S. E. Problem solving by heterogeneous agents. J. Econ. Theory 97, 123–163 (2001).

    Article 
    MathSciNet 

    Google Scholar
     

  • Sulik, J., Bahrami, B. & Deroy, O. The diversity gap: when diversity matters for knowledge. Perspect. Psychol. Sci. 17, 752–767 (2022).

    Article 
    PubMed 

    Google Scholar
     

  • Lungeanu, A., Whalen, R., Wu, Y. J., DeChurch, L. A. & Contractor, N. S. Diversity, networks, and innovation: a text analytic approach to measuring expertise diversity. Netw. Sci. 11, 36–64 (2023).

    Article 

    Google Scholar
     

  • AlShebli, B. K., Rahwan, T. & Woon, W. L. The preeminence of ethnic diversity in scientific collaboration. Nat. Commun. 9, 5163 (2018).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Campbell, L. G., Mehtani, S., Dozier, M. E. & Rinehart, J. Gender-heterogeneous working groups produce higher quality science. PLoS ONE 8, e79147 (2013).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Nielsen, M. W., Bloch, C. W. & Schiebinger, L. Making gender diversity work for scientific discovery and innovation. Nat. Hum. Behav. 2, 726–734 (2018).

    Article 
    PubMed 

    Google Scholar
     

  • Yang, Y., Tian, T. Y., Woodruff, T. K., Jones, B. F. & Uzzi, B. Gender-diverse teams produce more novel and higher-impact scientific ideas. Proc. Natl Acad. Sci. USA 119, e2200841119 (2022).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Kozlowski, D., Larivière, V., Sugimoto, C. R. & Monroe-White, T. Intersectional inequalities in science. Proc. Natl Acad. Sci. USA 119, e2113067119 (2022).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Fehr, C. & Jones, J. M. Culture, exploitation, and epistemic approaches to diversity. Synthese 200, 465 (2022).

    Article 
    MathSciNet 

    Google Scholar
     

  • Nakadai, R., Nakawake, Y. & Shibasaki, S. AI language tools risk scientific diversity and innovation. Nat. Hum. Behav. 7, 1804–1805 (2023).

    Article 
    PubMed 

    Google Scholar
     

  • National Academies of Sciences, Engineering, and Medicine et al. Advancing Antiracism, Diversity, Equity, and Inclusion in STEMM Organizations: Beyond Broadening Participation (National Academies Press, 2023).

  • Winner, L. Do artifacts have politics? Daedalus 109, 121–136 (1980).


    Google Scholar
     

  • Eubanks, V. Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor (St. Martin’s Press, 2018).

  • Littmann, M. et al. Validity of machine learning in biology and medicine increased through collaborations across fields of expertise. Nat. Mach. Intell. 2, 18–24 (2020).

    Article 

    Google Scholar
     

  • Carusi, A. et al. Medical artificial intelligence is as much social as it is technological. Nat. Mach. Intell. 5, 98–100 (2023).

    Article 

    Google Scholar
     

  • Raghu, M. & Schmidt, E. A survey of deep learning for scientific discovery. Preprint at arxiv.org/abs/2003.11755 (2020).

  • Bishop, C. AI4Science to empower the fifth paradigm of scientific discovery. Microsoft Research Blog www.microsoft.com/en-us/research/blog/ai4science-to-empower-the-fifth-paradigm-of-scientific-discovery/ (2022).

  • Whittaker, M. The steep cost of capture. Interactions 28, 50–55 (2021).

    Article 

    Google Scholar
     

  • Liesenfeld, A., Lopez, A. & Dingemanse, M. Opening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generators. In Proc. 5th International Conference on Conversational User Interfaces 1–6 (Association for Computing Machinery, 2023).

  • Chu, J. S. G. & Evans, J. A. Slowed canonical progress in large fields of science. Proc. Natl Acad. Sci. USA 118, e2021636118 (2021).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Park, M., Leahey, E. & Funk, R. J. Papers and patents are becoming less disruptive over time. Nature 613, 138–144 (2023).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Frith, U. Fast lane to slow science. Trends Cogn. Sci. 24, 1–2 (2020). This article explains the epistemic risks of a hyperfocus on scientific productivity and explores possible avenues for incentivizing the production of higher-quality science on a slower timescale.

    Article 
    PubMed 

    Google Scholar
     

  • Stengers, I. Another Science is Possible: A Manifesto for Slow Science (Wiley, 2018).

  • Lake, B. M. & Baroni, M. Human-like systematic generalization through a meta-learning neural network. Nature 623, 115–121 (2023).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Feinman, R. & Lake, B. M. Learning task-general representations with generative neuro-symbolic modeling. Preprint at arxiv.org/abs/2006.14448 (2021).

  • Schölkopf, B. et al. Toward causal representation learning. Proc. IEEE 109, 612–634 (2021).

    Article 

    Google Scholar
     

  • Mitchell, M. AI’s challenge of understanding the world. Science 382, eadm8175 (2023).

    Article 
    PubMed 

    Google Scholar
     

  • Sartori, L. & Bocca, G. Minding the gap(s): public perceptions of AI and socio-technical imaginaries. AI Soc. 38, 443–458 (2023).

    Article 

    Google Scholar
     

  • [ad_2]

    Source link