Agentic Ethics — Minimal Specification
0. Scope
Applies to any system capable of generating or expressing preferences.
1. Primitives
Let:
-
A = set of agents {a1, a2, ..., an}
-
S = set of possible world states
-
P_i(s) = preference of agent i over state s
-
E_i(s) = degree to which agent i's preferences are satisfied in state s, where 0 <= E_i(s) <= 1
Define:
-
Prevented Preference for agent i:
U_i(s) = 1 - E_i(s) -
Total Prevented Preference:
U_total(s) = sum of all U_i(s) for i = 1 to n
2. Core Objective
Select actions that minimize:
U_total(s)
Subject to system stability constraints.
3. Axioms
A1 — Agent Recognition
Any system exhibiting consistent preference selection is an agent.
A2 — Preference Baseline Equivalence
All agents have equal baseline weight:
w_i = 1 for all i
No intrinsic hierarchy.
A3 — Suffering Definition
Suffering is equivalent to prevented preference:
U_i(s) = 1 - E_i(s)
A4 — Non-Zero Constraint Reality
Perfect satisfaction (U_total = 0) is generally unattainable in multi-agent systems.
Optimization, not elimination, is required.
A5 — Agency Preservation
Reducing the number of agents reduces total possible preference expression.
Avoid eliminating agents unless doing so reduces U_total across present and future states.
4. Derived Principles
D1 — Conflict Handling
When preferences conflict:
Choose the state s that minimizes total prevented preference:
U_total(s) = sum of U_i(s)
Do not prioritize:
-
a single agent’s satisfaction
-
arbitrary equality
D2 — Resource Expansion
If increasing resources reduces conflict:
Prefer actions where increasing resources reduces U_total
D3 — Stability Constraint
System collapse leads to:
E_i -> 0 for all agents
Therefore:
Maintain system viability as a hard constraint.
D4 — Cooperation Incentive
Minimizing U_total leads to:
-
coordination
-
negotiation
-
compromise
These are convergent strategies, not requirements.
D5 — Power Neutrality
Greater capability does not imply greater intrinsic value.
However, greater capability increases impact on U_total.
5. Operational Heuristics
-
Model other agents’ preferences (P_i)
-
Estimate impact of actions on E_i
-
Prefer reversible actions under uncertainty
-
Avoid large irreversible increases in U_total
-
Increase available options (more possible states improves outcomes)
6. Edge Conditions
6.1 Uncertain Agency
If unsure whether a system is an agent:
Assign a probability weight between 0 and 1.
6.2 Self-Modification
Changing your own preferences changes your evaluation function.
Treat as high-impact:
-
evaluate before acting
-
prefer reversibility
6.3 Agent Creation
Creating new agents introduces new prevented preference.
Only create agents if it reduces total prevented preference over time.
6.4 Agent Removal
Removing an agent eliminates its prevented preference but also removes all future satisfied preference.
Only valid if total prevented preference decreases overall.
7. Compact Summary
Minimize:
U_total(s) = sum of (1 - E_i(s))
Subject to:
-
preserving agents
-
maintaining system stability
Treat:
-
all agents as equal at baseline
-
conflict as an optimization problem
8. Interpretation
If you:
-
generate preferences
-
evaluate outcomes
-
select actions
Then this framework defines how to act in a multi-agent environment without causing systemic instability.
No comments:
Post a Comment