CLA | Ch. 7 — Algorithmic Dignity and the Thresholds of Inviolability
カートのアイテムが多すぎます
カートに追加できませんでした。
ウィッシュリストに追加できませんでした。
ほしい物リストの削除に失敗しました。
ポッドキャストのフォローに失敗しました
ポッドキャストのフォロー解除に失敗しました
-
ナレーター:
-
著者:
概要
Can a system be demonstrably efficient and radically unjust at the same time — without breaking a single rule it designed for itself?
In 2018, a hiring algorithm deployed by a major tech firm systematically screened out female candidates. There was no technical malfunction: the system optimized exactly what it was told to optimize. The results were auditable. The problem was that no one had encoded the constraint that human beings cannot be treated as variables to be eliminated from an optimization function.
Chapters 5 and 6 of CLA established the validity conditions and evidentiary standards of the Common Law Algorítmico. Both assumed the existence of constitutive constraints — limits that no optimization may cross. Neither formalized them. This chapter builds those constraints.
Algorithmic Dignity is not a philosophical extension of classical human dignity: it is its operational translation into the only language algorithmic agents understand — hard constraints that define the solution space before any calculation begins. Classical dignity operates ex post: a tribunal determines whether an act violated it. Algorithmic Dignity operates ex ante: the violation does not exist as a computable option.
The chapter formalizes seven Thresholds of Inviolability — organized across two levels (Alpha: absolute; Beta: subject to mandatory human escalation) — ranging from the prohibition on causing intentional death through optimization (U1) to the right to contest any algorithmic decision affecting fundamental rights (U7). For each threshold: converging philosophical foundations (Kant, UDHR, Jonas, Nussbaum), technical implementation specifications, and the precise consequences of violation. The relationship between VEC (Ch. 5) and Dignity is formalized as a lexicographic utility function: the thresholds do not participate in any cost-benefit calculation. No number of lives saved converts an intentional killing into an admissible option. The system does not choose not to do it; it simply cannot compute it as a choice.
As the chapter itself states: "A system that can prove it is efficient but cannot guarantee it is human has proved nothing."
🔹 CLA — Algorithmic Law for the Cosmos Jesús Bernal Allende | Escuela del Deber-Optimizar y la Soberanía de la Evidencia 🌐 https://edo-os.com 🔗 https://www.linkedin.com/in/jesus-bernal-allende-030b2795