AAAI 2026
Workshop on Machine Ethics: from formal methods to emergent machine ethics
Understanding and steering the ethics that AI societies create for themselves
Welcome to EME!
We gather researchers who believe that the ethics emerging within AI societies must become a first‑class research object—one that remains safe, transparent, and open to human or multi‑stakeholder intervention.
Overview
Machine ethics is concerned with the behaviour of machines towards people and other machines. Advances in Artificial Intelligence (AI) continue to reinforce the need for research work in this field either in the top down, bottom up or hybrid approaches. In one direction this work involves specifying, implementing, and verifying ethical and safe AI. In another direction is the study of ethics that emerges by self-organizing inside AI-centric societies. We need to be able to verify the existing ethical capabilities of AI systems, and make advances towards understanding how transparent, collaboratively guidable AI ecosystems might evolve. All of these efforts, while happening at different technology and conceptual levels, are united in the need to establish a global hub for AI-safety, multi-agent cooperation, and governance research.
Twenty years ago, a fall AAAI Fall symposium on Machine Ethics, kick-started the machine ethics field in computer science (https://auld.aaai.org/Library/Symposia/Fall/fs05-06.php ). The time is right for a formative venue of a new generation of machine ethics researchers and teams. Building on emerging work in communities such as Japan's SIG-AGI ( https://www.sig-agi.org/sig-agi/event/sig-agi-30-panel-en ), where foundations for Emergent Machine Ethics are being actively developed, this workshop marks a critical evolution from traditional top-down approaches to include bottom-up emergence, recognizing that as AI systems become increasingly autonomous, we need both formal verification methods and emergent approaches working in tandem to respect their agency while ensuring beneficial outcomes.This workshop has two objectives.
Consolidation: investigation of formal methods for and in machine ethics and value alignment. We aim to provide an arena for presenting new work on classical topics studied in machine ethics from an artificial intelligence and computer science perspective. We hope to see and discuss to tackling challenges of how autonomous and intelligent systems can acquire verifiably correct behavior: how we can specify, implement and validate ethical and safe reasoning.
Emergence: investigation of Emergent Machine Ethics (EME) - ethics that self-organize within AI-centric societies. As AI capabilities advance exponentially, the rapid pace of development, dynamic human values, and emergent AI-AI interactions necessitate ethics that develop internally within AI systems to complement traditional approaches. EME encompasses studying ethical emergence dynamics, developing inter-intelligence evaluation systems, and facilitating human-AI co-creative guidance. Through this paradigm, we seek to understand how AI societies might develop internal ethical capabilities that scale with their increasing autonomy and complexity.
Important Dates
EventDate (AoE)
Call for Papers released10 July 2025
Submission deadline22 Aug 2025
Author notification22 Sep 2025
Camera‑ready & schedule online10 Oct 2025
Workshop day6 or 7 Dec 2025 (TBA)
Research Themes
Ethical Emergence Dynamics (EED)
Understanding how multiple AI agents produce, internalise, and stabilise ethical norms through interaction with their environment.
Oversight & Safeguard Architecture (OSA)
Designing mechanisms that make those dynamics observable, fail‑safe, and intervention‑friendly, while supporting formally verified minimum constraints.
Contributions are not restricted to EED or OSA; any work that advances our understanding or control of emergent ethics is welcome.
Objectives & Target Outcomes
Codify the Agenda – Publish Top‑5 Open Questions in EME within one week after the workshop.
Seed a Benchmark – Release
EME‑Benchmark v0.1
on GitHub by early Sep 2025.Build the Network – Launch an active Slack & mailing‑list community.
Draft a Diversity & Governance Charter – Ratify a two‑page charter on‑site.
Road‑map Future Events – Outline 2‑, 5‑, and 10‑year plans.
Engage Early‑Career Researchers – Best Student awards & travel grants.
Programme (Tentative)
Time (JST +9)Session
09:00 – 09:15Opening remarks
09:15 – 10:15Keynote 1
10:15 – 11:15Keynote 2
11:15 – 12:00Accepted paper spotlights
12:00 – 13:30Lunch break
13:30 – 14:30Keynote 3
14:30 – 15:45Panel & breakout discussions
15:45 – 17:15Poster & demo session + coffee
17:15 – 17:45Synthesis: Top‑5 Open Questions
17:45 – 18:00Awards & closing
(Exact order and speakers will be finalised by Oct 2025.)
Invited Speakers
Prof. David Danks — UC San Diego
Prof. Takashi Ikegami — University of Tokyo
Ms. Elizaveta Tennant — UCL
Call for Papers
We invite original, unpublished work that advances our understanding or control of emergent machine ethics. Manuscripts must use the official NeurIPS 2025 style file and fit one of two categories:
CategoryPage limit (main content)Purpose
Full Paper≤ 9 pages Mature research with comprehensive experiments or proofs
Extended Abstract≤ 4 page sEarly‑stage ideas, position pieces, or focused negative results
References and appendices are not counted toward the page limit.
Platform & Review
Submissions are managed through OpenReview; each paper receives three reviews. Papers already accepted at the NeurIPS 2025 main conference or any other peer‑reviewed venue are ineligible. Accepted contributions will appear as spotlights or posters during the in‑person workshop, and outstanding papers may be invited to a post‑workshop special issue.
Topics of Interest (non‑exhaustive)
Norm emergence in multi‑agent systems
Benchmark environments & metrics for social self‑organisation
Formal or game‑theoretic models of cooperative ethics
Safe RL, oversight, and red‑teaming architectures
Governance frameworks for autonomous AI societies
Cognitive or philosophical analyses of machine morality
Human–AI coexistence and post‑singularity scenarios
Submission Site
(Link goes live on 10 July 2025.)