Publications by year
In Press
Obradovich N, Özak Ö, Martín I, Ortuno-Ortin I, Awad E, Cebrián M, Cuevas R, Desmet K, Rahwan I, Cuevas Á, et al (In Press). Expanding the Measurement of Culture with a Sample of Two Billion Humans.
SSRN Electronic Journal DOI.
2023
Zhang Y, Ling S, Awad E, Frank MR, Du N (2023). Driving Next to Automated Vehicles: Emergent Human-machine Cooperation in Mixed Traffic. CHI '23: CHI Conference on Human Factors in Computing Systems.
DOI.
Franklin M, Awad E, Ashton H, Lagnado D (2023). Unpredictable robots elicit responsibility attributions.
Behav Brain Sci,
46Abstract:
Unpredictable robots elicit responsibility attributions.
Do people hold robots responsible for their actions? While Clark and Fischer present a useful framework for interpreting social robots, we argue that they fail to account for people's willingness to assign responsibility to robots in certain contexts, such as when a robot performs actions not predictable by its user or programmer.
Abstract.
Author URL.
DOI.
2022
Franklin M, Ashton H, Awad E, Lagnado D (2022). Causal Framework of Artificial Autonomous Agent Responsibility.
Abstract:
Causal Framework of Artificial Autonomous Agent Responsibility
Abstract.
DOI.
Jamison J, Awad E (2022). Computational ethics.
Trends in Cognitive Sciences DOI.
Obradovich N, Özak Ö, Martín I, Ortuño-Ortín I, Awad E, Cebrián M, Cuevas R, Desmet K, Rahwan I, Cuevas Á, et al (2022). Expanding the measurement of culture with a sample of two billion humans.
Journal of the Royal Society Interface,
19(190).
Abstract:
Expanding the measurement of culture with a sample of two billion humans
Culture has played a pivotal role in human evolution. Yet, the ability of social scientists to study culture is limited by the currently available measurement instruments. Scholars of culture must regularly choose between scalable but sparse survey-based methods or restricted but rich ethnographic methods. Here, we demonstrate that massive online social networks can advance the study of human culture by providing quantitative, scalable and high-resolution measurement of behaviourally revealed cultural values and preferences. We employ data across nearly 60 000 topic dimensions drawn from two billion Facebook users across 225 countries and territories. We first validate that cultural distances calculated from this measurement instrument correspond to traditional survey-based and objective measures of cross-national cultural differences. We then demonstrate that this expanded measure enables rich insight into the cultural landscape globally at previously impossible resolution. We analyse the importance of national borders in shaping culture and compare subnational divisiveness with gender divisiveness across countries. Our measure enables detailed investigation into the geopolitical stability of countries, social cleavages within small- and large-scale human groups, the integration of migrant populations and the disaffection of certain population groups from the political process, among myriad other potential future applications.
Abstract.
DOI.
Pick CM, Ko A, Kenrick DT, Wiezel A, Wormley AS, Awad E, Al-Shawaf L, Barry O, Bereby-Meyer Y, Boonyasiriwat W, et al (2022). Fundamental social motives measured across forty-two cultures in two waves.
Scientific Data,
9(1).
Abstract:
Fundamental social motives measured across forty-two cultures in two waves
AbstractHow does psychology vary across human societies? the fundamental social motives framework adopts an evolutionary approach to capture the broad range of human social goals within a taxonomy of ancestrally recurring threats and opportunities. These motives—self-protection, disease avoidance, affiliation, status, mate acquisition, mate retention, and kin care—are high in fitness relevance and everyday salience, yet understudied cross-culturally. Here, we gathered data on these motives in 42 countries (N = 15,915) in two cross-sectional waves, including 19 countries (N = 10,907) for which data were gathered in both waves. Wave 1 was collected from mid-2016 through late 2019 (32 countries, N = 8,998; 3,302 male, 5,585 female; Mage = 24.43, SD = 7.91). Wave 2 was collected from April through November 2020, during the COVID-19 pandemic (29 countries, N = 6,917; 2,249 male, 4,218 female; Mage = 28.59, SD = 11.31). These data can be used to assess differences and similarities in people’s fundamental social motives both across and within cultures, at different time points, and in relation to other commonly studied cultural indicators and outcomes.
Abstract.
DOI.
Awad E, Bago B, Bonnefon J-F, Christakis NA, Rahwan I, Shariff A (2022). Polarized Citizen Preferences for the Ethical Allocation of Scarce Medical Resources in 20 Countries.
MDM Policy Pract,
7(2).
Abstract:
Polarized Citizen Preferences for the Ethical Allocation of Scarce Medical Resources in 20 Countries.
UNLABELLED: Objective. When medical resources are scarce, clinicians must make difficult triage decisions. When these decisions affect public trust and morale, as was the case during the COVID-19 pandemic, experts will benefit from knowing which triage metrics have citizen support. Design. We conducted an online survey in 20 countries, comparing support for 5 common metrics (prognosis, age, quality of life, past and future contribution as a health care worker) to a benchmark consisting of support for 2 no-triage mechanisms (first-come-first-served and random allocation). Results. We surveyed nationally representative samples of 1000 citizens in each of Brazil, France, Japan, and the United States and also self-selected samples from 20 countries (total N = 7599) obtained through a citizen science website (the Moral Machine). We computed the support for each metric by comparing its usability to the usability of the 2 no-triage mechanisms. We further analyzed the polarizing nature of each metric by considering its usability among participants who had a preference for no triage. In all countries, preferences were polarized, with the 2 largest groups preferring either no triage or extensive triage using all metrics. Prognosis was the least controversial metric. There was little support for giving priority to healthcare workers. Conclusions. It will be difficult to define triage guidelines that elicit public trust and approval. Given the importance of prognosis in triage protocols, it is reassuring that it is the least controversial metric. Experts will need to prepare strong arguments for other metrics if they wish to preserve public trust and morale during health crises. HIGHLIGHTS: We collected citizen preferences regarding triage decisions about scarce medical resources from 20 countries.We find that citizen preferences are universally polarized.Citizens either prefer no triage (random allocation or first-come-first served) or extensive triage using all common triage metrics, with "prognosis" being the least controversial.Experts will need to prepare strong arguments to preserve or elicit public trust in triage decisions.
Abstract.
Author URL.
DOI.
Pick CM, Ko A, Kenrick DT, Wiezel A, Wormley AS, Awad E, Al-Shawaf L, Barry O, Bereby-Meyer Y, Boonyasiriwat W, et al (2022). Publisher Correction: Fundamental social motives measured across forty-two cultures in two waves.
Scientific Data,
9(1).
DOI.
2021
Saveski M, Awad E, Rahwan I, Cebrian M (2021). Algorithmic and human prediction of success in human collaboration from visual features.
Sci Rep,
11(1).
Abstract:
Algorithmic and human prediction of success in human collaboration from visual features.
As groups are increasingly taking over individual experts in many tasks, it is ever more important to understand the determinants of group success. In this paper, we study the patterns of group success in Escape the Room, a physical adventure game in which a group is tasked with escaping a maze by collectively solving a series of puzzles. We investigate (1) the characteristics of successful groups, and (2) how accurately humans and machines can spot them from a group photo. The relationship between these two questions is based on the hypothesis that the characteristics of successful groups are encoded by features that can be spotted in their photo. We analyze >43K group photos (one photo per group) taken after groups have completed the game-from which all explicit performance-signaling information has been removed. First, we find that groups that are larger, older and more gender but less age diverse are significantly more likely to escape. Second, we compare humans and off-the-shelf machine learning algorithms at predicting whether a group escaped or not based on the completion photo. We find that individual guesses by humans achieve 58.3% accuracy, better than random, but worse than machines which display 71.6% accuracy. When humans are trained to guess by observing only four labeled photos, their accuracy increases to 64%. However, training humans on more labeled examples (eight or twelve) leads to a slight, but statistically insignificant improvement in accuracy (67.4%). Humans in the best training condition perform on par with two, but worse than three out of the five machine learning algorithms we evaluated. Our work illustrates the potentials and the limitations of machine learning systems in evaluating group performance and identifying success factors based on sparse visual cues.
Abstract.
Author URL.
DOI.
Franklin M, Awad E, Lagnado D (2021). Blaming automated vehicles in difficult situations.
iScience,
24Abstract:
Blaming automated vehicles in difficult situations
Automated Vehicles (AVs) have made huge strides towards large scale deployment.
Despite this progress, AVs continue to make mistakes, some resulting in death. While
some mistakes are avoidable, others are hard to avoid even by highly-skilled drivers. As
these mistakes continue to shape attitudes towards AVs, we need to understand whether
people differentiate between them. We ask the following two questions. When an AV
makes a mistake, does the perceived difficulty or novelty of the situation predict blame
attributed to it? How does that blame attribution compare to a human driving a car?
Through two studies we find that the amount of blame people attribute to AVs and
human drivers is sensitive to situation difficulty. However, while some situations could
be more difficult for AVs and others for human drivers, people blamed AVs more,
regardless. Our results provide novel insights in understanding psychological barriers
influencing the public’s view of AVs.
Abstract.
DOI.
Everett JAC, Colombatto C, Awad E, Boggio P, Bos B, Brady WJ, Chawla M, Chituc V, Chung D, Drupp MA, et al (2021). Moral dilemmas and trust in leaders during a global health crisis.
Nat Hum Behav,
5(8), 1074-1088.
Abstract:
Moral dilemmas and trust in leaders during a global health crisis.
Trust in leaders is central to citizen compliance with public policies. One potential determinant of trust is how leaders resolve conflicts between utilitarian and non-utilitarian ethical principles in moral dilemmas. Past research suggests that utilitarian responses to dilemmas can both erode and enhance trust in leaders: sacrificing some people to save many others ('instrumental harm') reduces trust, while maximizing the welfare of everyone equally ('impartial beneficence') may increase trust. In a multi-site experiment spanning 22 countries on six continents, participants (N = 23,929) completed self-report (N = 17,591) and behavioural (N = 12,638) measures of trust in leaders who endorsed utilitarian or non-utilitarian principles in dilemmas concerning the COVID-19 pandemic. Across both the self-report and behavioural measures, endorsement of instrumental harm decreased trust, while endorsement of impartial beneficence increased trust. These results show how support for different ethical principles can impact trust in leaders, and inform effective public communication during times of global crisis. PROTOCOL REGISTRATION STATEMENT: the Stage 1 protocol for this Registered Report was accepted in principle on 13 November 2020. The protocol, as accepted by the journal, can be found at https://doi.org/10.6084/m9.figshare.13247315.v1.
Abstract.
Author URL.
DOI.
2020
Awad E, Anderson M, Anderson SL, Liao B (2020). An approach for combining ethical principles with public opinion to guide public policy.
Artificial Intelligence,
287Abstract:
An approach for combining ethical principles with public opinion to guide public policy
We propose a framework for incorporating public opinion into policy making in situations where values are in conflict. This framework advocates creating vignettes representing value choices, eliciting the public's opinion on these choices, and using machine learning to extract principles that can serve as succinct statements of the policies implied by these choices and rules to guide the behavior of autonomous systems.
Abstract.
DOI.
Awad E, Dsouza S, Bonnefon JF, Shariff A, Rahwan I (2020). Crowdsourcing moral machines.
Communications of the ACM,
63(3), 48-55.
DOI.
Awad E, Dsouza S, Shariff A, Rahwan I, Bonnefon J-F (2020). Reply to Claessens et al.: Maybe the Footbridge sacrifice is indeed the only one that sends a negative social signal.
Proceedings of the National Academy of Sciences,
117(24), 13205-13206.
DOI.
Awad E, Dsouza S, Kim R, Schulz J, Henrich J, Shariff A, Bonnefon JF, Rahwan I (2020). Reply to: Life and death decisions of autonomous vehicles.
Nature,
579(7797), E3-E5.
DOI.
Awad E, Dsouza S, Shariff A, Rahwan I, Bonnefon J-F (2020). Universals and variations in moral decisions made in 42 countries by 70,000 participants.
Proc Natl Acad Sci U S A,
117(5), 2332-2337.
Abstract:
Universals and variations in moral decisions made in 42 countries by 70,000 participants.
When do people find it acceptable to sacrifice one life to save many? Cross-cultural studies suggested a complex pattern of universals and variations in the way people approach this question, but data were often based on small samples from a small number of countries outside of the Western world. Here we analyze responses to three sacrificial dilemmas by 70,000 participants in 10 languages and 42 countries. In every country, the three dilemmas displayed the same qualitative ordering of sacrifice acceptability, suggesting that this ordering is best explained by basic cognitive processes rather than cultural norms. The quantitative acceptability of each sacrifice, however, showed substantial country-level variations. We show that low relational mobility (where people are more cautious about not alienating their current social partners) is strongly associated with the rejection of sacrifices for the greater good (especially for Eastern countries), which may be explained by the signaling value of this rejection. We make our dataset fully available as a public resource for researchers studying universals and variations in human morality.
Abstract.
Author URL.
DOI.
2019
Awad E, Levine S, Kleiman-Weiner M, Dsouza S, Tenenbaum JB, Shariff A, Bonnefon J-F, Rahwan I (2019). Drivers are blamed more than their automated cars when both make mistakes.
Nature Human Behaviour,
3(10).
Abstract:
Drivers are blamed more than their automated cars when both make mistakes
When an automated car harms someone, who is blamed by those who hear about it? Here, we asked human participants to consider hypothetical cases in which a pedestrian was killed by a car operated under shared control of a primary and a secondary driver, and to indicate how blame should be allocated. We find that when only one driver makes an error, that driver is blamed more, regardless of whether that driver is a machine or a human. However, when both drivers make errors in cases of human-machine shared-control vehicles, the blame attributed to the machine is reduced. This finding portends a public under-reaction to the malfunctioning AI components of automated cars and therefore has a direct policy implication: allowing the de-facto standards for shared-control vehicles to be established in courts by the jury system could fail to properly regulate the safety of those vehicles; instead, a top-down scheme (through federal laws) may be called for.
Abstract.
DOI.
2018
Kim R, Kleiman-Weiner M, Abeliuk A, Awad E, Dsouza S, Tenenbaum JB, Rahwan I (2018). A Computational Model of Commonsense Moral Decision Making. AIES '18: AAAI/ACM Conference on AI, Ethics, and Society.
DOI.
Noothigattu R, Gaikwad SNS, Awad E, Dsouza S, Rahwan I, Ravikumar P, Procaccia AD (2018). A voting-based system for ethical decision making.
Abstract:
A voting-based system for ethical decision making
Abstract.
Awad E, Dsouza S, Kim R, Schulz J, Henrich J, Shariff A, Bonnefon J-F, Rahwan I (2018). The Moral Machine experiment.
Nature,
563 (7729), 59-64.
Abstract:
The Moral Machine experiment
With the rapid development of artificial intelligence have come concerns about how machines will make moral decisions, and the major challenge of quantifying societal expectations about the ethical principles that should guide machine behaviour. To address this challenge, we deployed the Moral Machine, an online experimental platform designed to explore the moral dilemmas faced by autonomous vehicles. This platform gathered 40 million decisions in ten languages from millions of people in 233 countries and territories. Here we describe the results of this experiment. First, we summarize global moral preferences. Second, we document individual variations in preferences, based on respondents’ demographics. Third, we report cross-cultural ethical variation, and uncover three major clusters of countries. Fourth, we show that these differences correlate with modern institutions and deep cultural traits. We discuss how these preferences can contribute to developing global, socially acceptable principles for machine ethics. All data used in this article are publicly available.
Abstract.
DOI.
2017
Awad E, Bonnefon J-F, Caminada M, Malone TW, Rahwan I (2017). Experimental Assessment of Aggregation Principles in Argumentation-Enabled Collective Intelligence.
ACM Transactions on Internet Technology,
17(3), 1-21.
Abstract:
Experimental Assessment of Aggregation Principles in Argumentation-Enabled Collective Intelligence
On the Web, there is always a need to aggregate opinions from the crowd (as in posts, social networks, forums, etc.). Different mechanisms have been implemented to capture these opinions such as like in Facebook, Favorite in Twitter, thumbs-up/-down, flagging, and so on. However, in more contested domains (e.g. Wikipedia, political discussion, and climate change discussion), these mechanisms are not sufficient, since they only deal with each issue independently without considering the relationships between different claims. We can view a set of conflicting arguments as a graph in which the nodes represent arguments and the arcs between these nodes represent the defeat relation. A group of people can then collectively evaluate such graphs. To do this, the group must use a rule to aggregate their individual opinions about the entire argument graph. Here we present the first experimental evaluation of different principles commonly employed by aggregation rules presented in the literature. We use randomized controlled experiments to investigate which principles people consider better at aggregating opinions under different conditions. Our analysis reveals a number of factors, not captured by traditional formal models, that play an important role in determining the efficacy of aggregation. These results help bring formal models of argumentation closer to real-world application.
Abstract.
DOI.
Awad E, Caminada MWA, Pigozzi G, Podlaszewski M, Rahwan I (2017). Pareto optimality and strategy-proofness in group argument evaluation.
Journal of Logic and Computation,
27(8), 2581-2609.
DOI.
2015
Awad E, Booth R, Tohmé F, Rahwan I (2015). Judgement aggregation in multi-agent argumentation.
Journal of Logic and Computation,
27(1), 227-259.
DOI.
Alshamsi A, Awad E, Almehrezi M, Babushkin V, Chang P-J, Shoroye Z, Tóth A-P, Rahwan I (2015). Misery loves company: happiness and communication in the city.
EPJ Data Science,
4(1).
DOI.
2014
Booth R, Awad E, Rahwan I (2014). Interval methods for judgment aggregation in argumentation.
Abstract:
Interval methods for judgment aggregation in argumentation
Abstract.