Newsletter

When AI Chooses Who Lives (and Who Dies): The Modern Trolley Problem

The trolley dilemma in the age of AI: when machines have to make ethical decisions, is human judgment really always superior? The debate is still open. Why the ethics of algorithms could be better than human ethics (or maybe not).

Imagine a runaway train heading toward five people. You can pull a lever to divert it onto another track, but there is only one person there. What would you do?

But wait: what if that person were a child and the five were elderly? What if someone offered you money to pull the lever? What if you couldn't see the situation clearly?

What is the Trolley Problem? Formulated by philosopher Philippa Foot in 1967, this thought experiment presents a seemingly simple dilemma: sacrifice one life to save five. But the variations are endless: the fat man to be pushed off the bridge, the doctor who could kill one healthy patient to save five with his organs, the judge who could convict an innocent person to stop a riot.

Each scenario tests our fundamental moral principles: when is it acceptable to cause harm in order to prevent greater harm?

This complexity is precisely what makes the ethics of artificial intelligence such a crucial challenge for our time.

The famous "trolley problem" is much more complex than it seems—and this complexity is precisely what makes the ethics of artificial intelligence such a crucial challenge for our time.

From the Philosophy Classroom to Algorithms

The trolley problem, formulated by philosopher Philippa Foot in 1967, was never intended to solve practical dilemmas. As theAlan Turing Institute, the original purpose was to demonstrate that thought experiments are, in essence, divorced from reality. Yet, in the age of AI, this paradox has taken on immediate relevance.

Why is this important now? Because for the first time in history, machines must make ethical decisions in real time—from autonomous cars navigating traffic to healthcare systems allocating limited resources.

Claude and the Constitutional AI Revolution

Anthropic, the company behind Claude, tackled this challenge with a revolutionary approach called Constitutional AI. Instead of relying solely on human feedback, Claude is trained on a "constitution" of explicit ethical principles, including elements of the Universal Declaration of Human Rights.

How does it work in practice?

  • Claude self-criticizes and revises his answers.
  • Use Reinforcement Learning from AI Feedback (RLAIF)
  • Maintains transparency regarding the principles that guide its decisions

Anempirical analysis of 700,000 conversations revealed that Claude expresses over 3,000 unique values, from professionalism to moral pluralism, adapting them to different contexts while maintaining ethical consistency.

The Real Challenges: When Theory Meets Practice

As brilliantly illustrated by the interactive project Absurd Trolley Problems by Neal Agarwal brilliantly illustrates, real-world ethical dilemmas are rarely binary and often absurd in their complexity. This insight is crucial to understanding the challenges of modern AI.

Recent research shows that the ethical dilemmas of AI go far beyond the classic trolley problem. The MultiTP project MultiTPproject, which tested 19 AI models in over 100 languages, found significant cultural variations in ethical alignment: models are more aligned with human preferences in English, Korean, and Chinese, but less so in Hindi and Somali.

The real challenges include:

  • Epistemic uncertainty: Acting without complete information
  • Cultural biases: Different values across cultures and communities
  • Distributed responsibility: Who is responsible for AI decisions?
  • Long-term consequences: Immediate vs. future effects

Human Ethics vs. AI Ethics: Different Paradigms, Not Necessarily Worse

An often overlooked aspect is that AI ethics may not simply be an imperfect version of human ethics, but a completely different paradigm—and in some cases, potentially more consistent.

The Case of "I, Robot": In the 2004 film, Detective Spooner (Will Smith) is suspicious of robots after being saved by one in a car accident, while a 12-year-old girl was left to drown. The robot explains its decision:

"I was the logical choice. I calculated that she had a 45% chance of survival. Sarah only had 11%. That was someone's child. 11% is more than enough."

This is exactly the kind of ethics that AI operates on today: algorithms that weigh probabilities, optimize results, and make decisions based on objective data rather than emotional insights or social biases. The scene illustrates a crucial point: AI operates with ethical principles that are different but not necessarily inferior to human ones:

  • Mathematical consistency: Algorithms apply criteria uniformly, without being influenced by emotional or social biases—just like a robot calculating survival probabilities.
  • Procedural impartiality: They do not automatically favor children over the elderly or the rich over the poor, but evaluate each situation based on the available data.
  • Transparency in decision-making: The criteria are explicit and verifiable ("45% vs. 11%"), unlike human moral intuition, which is often opaque.

Concrete examples in modern AI:

  • AI healthcare systems that allocate medical resources based on the probability of therapeutic success
  • Matching algorithms for organ transplants that optimize compatibility and survival rates
  • Automated triage systems in emergencies that prioritize patients with the best chance of recovery

But Maybe Not: The Fatal Limits of Algorithmic Ethics

However, before celebrating the superiority of AI ethics, we must confront its inherent limitations. The scene from "I, Robot" that seems so logical hides profound problems:

The Problem of Lost Context: When the robot chooses to save the adult instead of the child based on probabilities, it completely ignores crucial elements:

  • The social and symbolic value of protecting the most vulnerable
  • The long-term psychological impact on survivors
  • Family relationships and emotional ties
  • The untapped potential of a young life

The Concrete Risks of Purely Algorithmic Ethics:

Extreme Reductionism: Turning complex moral decisions into mathematical calculations can remove human dignity from the equation. Who decides which variables matter?

Hidden Biases: Algorithms inevitably incorporate the biases of their creators and training data. A system that "optimizes" could perpetuate systemic discrimination.

Cultural Uniformity: AI ethics risks imposing a Western, technological, and quantitative view of morality on cultures that value human relationships differently.

Examples of real challenges:

  • Healthcare systems that could apply efficiency criteria more systematically, raising questions about how to balance medical optimization and ethical considerations
  • Judicial algorithms that risk perpetuating existing biases on a larger scale, but which could also make existing discrimination more transparent
  • Financial AI that can systematize discriminatory decisions, but also eliminate certain human biases linked to personal prejudices

Criticism of the Traditional Paradigm

Experts such as Roger Scruton criticize the use of the trolley problem for its tendency to reduce complex dilemmas to "pure arithmetic," eliminating morally relevant relationships. As argued in an article in TripleTen, "solving the trolley problem will not make AI ethical"—a more holistic approach is needed.

The central question becomes: Can we afford to delegate moral decisions to systems that, however sophisticated, lack empathy, contextual understanding, and human experiential wisdom?

New proposals for balance:

  • Hybrid ethical frameworks that combine computation and human intuition
  • Human oversight systems for critical decisions
  • Cultural customization of ethical algorithms
  • Mandatory transparency on decision-making criteria
  • Human right of appeal for all critical algorithmic decisions

Practical Implications for Companies

For business leaders, this evolution requires a nuanced approach:

  1. Systematic ethical audits of AI systems in use – to understand both advantages and limitations
  2. Diversity in teams that design and implement AI, including philosophers, ethicists, and representatives from diverse communities
  3. Mandatory transparency on the ethical principles incorporated into systems and their rationale
  4. Ongoing training on when AI ethics works and when it fails
  5. Human oversight systems for decisions with high ethical impact
  6. Appeal rights and correction mechanisms for algorithmic decisions

As highlighted by IBM in its 2025 outlook, AI literacy and clear accountability will be the most critical challenges for the coming year.

The Future of AI Ethics

UNESCO is leadingUNESCO is leading global initiatives for AI ethics, with the 3rd Global Forum scheduled for June 2025 in Bangkok. The goal is not to find universal solutions to moral dilemmas, but to develop frameworks that enable transparent and culturally sensitive ethical decisions.

The key lesson? The trolley problem serves not as a solution, but as a reminder of the inherent complexity of moral decisions. The real challenge is not choosing between human or algorithmic ethics, but finding the right balance between computational efficiency and human wisdom.

The ethical AI of the future will have to recognize its own limitations: excellent at processing data and identifying patterns, but inadequate when empathy, cultural understanding, and contextual judgment are required. As in the scene from "I, Robot," cold calculation can sometimes be more ethical—but only if it remains a tool in the hands of conscious human supervision, not a substitute for human moral judgment.

The "(or perhaps not)" in our title is not indecision, but wisdom: recognizing that ethics, whether human or artificial, does not allow for simple solutions in a complex world.

Sources and Insights

Initial Inspiration:

Academic Research:

Industrial Analysis:

Regulatory Developments:

Resources for business growth

November 9, 2025

Regulating what is not created: does Europe risk technological irrelevance?

Europe attracts only one-tenth of global investment in artificial intelligence but claims to dictate global rules. This is the "Brussels Effect"-imposing regulations on a planetary scale through market power without driving innovation. The AI Act goes into effect on a staggered timetable until 2027, but multinational tech companies respond with creative evasion strategies: invoking trade secrets to avoid revealing training data, producing technically compliant but incomprehensible summaries, using self-assessment to downgrade systems from "high risk" to "minimal risk," forum shopping by choosing member states with less stringent controls. The extraterritorial copyright paradox: EU demands that OpenAI comply with European laws even for training outside Europe-principle never before seen in international law. The "dual model" emerges: limited European versions vs. advanced global versions of the same AI products. Real risk: Europe becomes "digital fortress" isolated from global innovation, with European citizens accessing inferior technologies. The Court of Justice in the credit scoring case has already rejected the "trade secrets" defense, but interpretive uncertainty remains huge-what exactly does "sufficiently detailed summary" mean? No one knows. Final unresolved question: is the EU creating an ethical third way between U.S. capitalism and Chinese state control, or simply exporting bureaucracy to an industry where it does not compete? For now: world leader in AI regulation, marginal in its development. Vaste program.
November 9, 2025

Outliers: Where Data Science Meets Success Stories.

Data science has turned the paradigm on its head: outliers are no longer "errors to be eliminated" but valuable information to be understood. A single outlier can completely distort a linear regression model-change the slope from 2 to 10-but eliminating it could mean losing the most important signal in the dataset. Machine learning introduces sophisticated tools: Isolation Forest isolates outliers by building random decision trees, Local Outlier Factor analyzes local density, Autoencoders reconstruct normal data and report what they cannot reproduce. There are global outliers (temperature -10°C in tropics), contextual outliers (spending €1,000 in poor neighborhood), collective outliers (synchronized spikes traffic network indicating attack). Parallel with Gladwell: the "10,000 hour rule" is disputed-Paul McCartney dixit "many bands have done 10,000 hours in Hamburg without success, theory not infallible." Asian math success is not genetic but cultural: Chinese number system more intuitive, rice cultivation requires constant improvement vs Western agriculture territorial expansion. Real applications: UK banks recover 18% potential losses via real-time anomaly detection, manufacturing detects microscopic defects that human inspection would miss, healthcare valid clinical trials data with 85%+ sensitivity anomaly detection. Final lesson: as data science moves from eliminating outliers to understanding them, we must see unconventional careers not as anomalies to be corrected but as valuable trajectories to be studied.