🧠

EthicaAI

Beyond Homo Economicus: Computational Verification of Sen's Meta-Ranking Theory via Multi-Agent Reinforcement Learning

📄 Preprint 2026 🔬 7 SVO Conditions 🌍 3 Environments 👥 100 Agents
19
Figures
5
Key Findings
33
References
87%
Convergence
12%
ESS

⚙️ Settings

Prosocial (45°)
10
λ_base = sin(45°) = 0.707
Mode: Dynamic

📊 Contribution Rate

💰 Mean Payoff

Metric 1
Metric 2
Metric 3

Abstract

This study formalizes Amartya Sen's theory of "Meta-Ranking"—preferences over preferences—within a Multi-Agent Reinforcement Learning (MARL) framework. We simulated agents with seven Social Value Orientations (SVO) across three environments (Cleanup, IPD, PGG) at scales up to 100 agents.

Key Equation

R_total = (1 − λ_t) · U_self + λ_t · [U_meta − ψ]

where λ_t dynamically modulates between self-interest and social commitment based on resource levels, implementing Sen's insight that commitment is impossible under extreme deprivation.

Five Key Findings

# Finding Evidence
1 Dynamic meta-ranking enhances collective welfare p=0.0003
2 Emergent role specialization (Cleaners vs Eaters) p<0.0001 at 100 agents
3 "Situational Commitment" → ESS at ~12% Replicator dynamics
4 Individualist SVO (15°) best matches humans WD=0.053
5 SVO rotation = 86% of mechanism Full factorial

Three Implications

Domain Implication
🤖 AI Alignment Systems should learn when to be moral, not encode static values
📊 Behavioral Economics Bounded self-interest (θ=15°), not pure altruism, best replicates humans
🧬 Evolutionary Theory A "Moral Minority" of ~12% suffices as an ESS

Links

📁 GitHub Repository
📄 Zenodo DOI