AAAI 2026

Workshop on Machine Ethics: from formal methods to emergent machine ethics

Understanding and steering the ethics that AI societies create for themselves

Welcome to EME!
We gather researchers who believe that the ethics emerging within AI societies must become a first‑class research object—one that remains safe, transparent, and open to human or multi‑stakeholder intervention.

Overview

Machine ethics is concerned with the behaviour of machines towards people and other machines. Advances in Artificial Intelligence (AI) continue to reinforce the need for research work in this field either in the top down, bottom up or hybrid approaches. In one direction this work involves specifying, implementing, and verifying ethical and safe AI. In another direction is the study of ethics that emerges by self-organizing inside AI-centric societies. We need to be able to verify the existing ethical capabilities of AI systems, and make advances towards understanding how transparent, collaboratively guidable AI ecosystems might evolve. All of these efforts, while happening at different technology and conceptual levels, are united in the need to establish a global hub for AI-safety, multi-agent cooperation, and governance research.

Twenty years ago, a fall AAAI Fall symposium on Machine Ethics, kick-started the machine ethics field in computer science (https://auld.aaai.org/Library/Symposia/Fall/fs05-06.php ). The time is right for a formative venue of a new generation of machine ethics researchers and teams. Building on emerging work in communities such as Japan's SIG-AGI ( https://www.sig-agi.org/sig-agi/event/sig-agi-30-panel-en ), where foundations for Emergent Machine Ethics are being actively developed, this workshop marks a critical evolution from traditional top-down approaches to include bottom-up emergence, recognizing that as AI systems become increasingly autonomous, we need both formal verification methods and emergent approaches working in tandem to respect their agency while ensuring beneficial outcomes.This workshop has two objectives.

Consolidation: investigation of formal methods for and in machine ethics and value alignment. We aim to provide an arena for presenting new work on classical topics studied in machine ethics from an artificial intelligence and computer science perspective. We hope to see and discuss to tackling challenges of how autonomous and intelligent systems can acquire verifiably correct behavior: how we can specify, implement and validate ethical and safe reasoning.

Emergence: investigation of Emergent Machine Ethics (EME) - ethics that self-organize within AI-centric societies. As AI capabilities advance exponentially, the rapid pace of development, dynamic human values, and emergent AI-AI interactions necessitate ethics that develop internally within AI systems to complement traditional approaches. EME encompasses studying ethical emergence dynamics, developing inter-intelligence evaluation systems, and facilitating human-AI co-creative guidance. Through this paradigm, we seek to understand how AI societies might develop internal ethical capabilities that scale with their increasing autonomy and complexity.

Important Dates

EventDate (AoE)

  • Call for Papers released10 July 2025

  • Submission deadline22 Aug 2025

  • Author notification22 Sep 2025

  • Camera‑ready & schedule online10 Oct 2025

  • Workshop day6 or 7 Dec 2025 (TBA)

Research Themes

Ethical Emergence Dynamics (EED)

Understanding how multiple AI agents produce, internalise, and stabilise ethical norms through interaction with their environment.

Oversight & Safeguard Architecture (OSA)

Designing mechanisms that make those dynamics observable, fail‑safe, and intervention‑friendly, while supporting formally verified minimum constraints.

Contributions are not restricted to EED or OSA; any work that advances our understanding or control of emergent ethics is welcome.

Objectives & Target Outcomes

  1. Codify the Agenda – Publish Top‑5 Open Questions in EME within one week after the workshop.

  2. Seed a Benchmark – Release EME‑Benchmark v0.1 on GitHub by early Sep 2025.

  3. Build the Network – Launch an active Slack & mailing‑list community.

  4. Draft a Diversity & Governance Charter – Ratify a two‑page charter on‑site.

  5. Road‑map Future Events – Outline 2‑, 5‑, and 10‑year plans.

  6. Engage Early‑Career Researchers – Best Student awards & travel grants.

Programme (Tentative)

Time (JST +9)Session

  • 09:00 – 09:15Opening remarks

  • 09:15 – 10:15Keynote 1

  • 10:15 – 11:15Keynote 2

  • 11:15 – 12:00Accepted paper spotlights

  • 12:00 – 13:30Lunch break

  • 13:30 – 14:30Keynote 3

  • 14:30 – 15:45Panel & breakout discussions

  • 15:45 – 17:15Poster & demo session + coffee

  • 17:15 – 17:45Synthesis: Top‑5 Open Questions

  • 17:45 – 18:00Awards & closing

(Exact order and speakers will be finalised by Oct 2025.)

Invited Speakers

Prof. David Danks — UC San Diego

Prof. Takashi Ikegami — University of Tokyo

Ms. Elizaveta Tennant — UCL

Call for Papers

We invite original, unpublished work that advances our understanding or control of emergent machine ethics. Manuscripts must use the official NeurIPS 2025 style file and fit one of two categories:

CategoryPage limit (main content)Purpose

  • Full Paper≤ 9 pages Mature research with comprehensive experiments or proofs

  • Extended Abstract≤ 4 page sEarly‑stage ideas, position pieces, or focused negative results

References and appendices are not counted toward the page limit.

Platform & Review

Submissions are managed through OpenReview; each paper receives three reviews. Papers already accepted at the NeurIPS 2025 main conference or any other peer‑reviewed venue are ineligible. Accepted contributions will appear as spotlights or posters during the in‑person workshop, and outstanding papers may be invited to a post‑workshop special issue.

Topics of Interest (non‑exhaustive)

  • Norm emergence in multi‑agent systems

  • Benchmark environments & metrics for social self‑organisation

  • Formal or game‑theoretic models of cooperative ethics

  • Safe RL, oversight, and red‑teaming architectures

  • Governance frameworks for autonomous AI societies

  • Cognitive or philosophical analyses of machine morality

  • Human–AI coexistence and post‑singularity scenarios

Submission Site

(Link goes live on 10 July 2025.)