Complex Value Systems are Required to Realize Valuable Futures – A Critique
|

Complex Value Systems are Required to Realize Valuable Futures – A Critique

In his 2011 paper, Complex Value Systems are Required to Realize Valuable Futures , Eliezer Yudkowsky posits that aligning artificial general intelligence (AGI) or artificial superintelligence (ASI) with human values necessitates embedding the complex intricacies of human ethics into AI systems. He warns against oversimplification, suggesting that without a comprehensive inheritance of human values, AGI…

AI Values: Satiable vs Insatiable
| |

AI Values: Satiable vs Insatiable

In short: Satiable values have diminishing returns with more resources, while insatiable values always want more, potentially leading to risky AI behaviour – it seems likely that AI with insatiable values could make existential trades, like gambling the world for a chance to double resources. Therefore potentially design AI with satiable values to ensure stability,…

On the Emergence of Biased Coherent Value Systems in AI as Value Risk
|

On the Emergence of Biased Coherent Value Systems in AI as Value Risk

Firstly, values are important because they are the fundamental principles that guide decisions and actions in humans. They shape our understanding of right and wrong, and they influence how we interact with the world around us. In the context of AI, values are particularly important because they determine how AI systems will behave, what goals…

The Architecture of Value
| |

The Architecture of Value

Implications for AI Safety and Indirect Normativity The challenge of understanding value space has moved from a philosophical curiosity to an urgent technical problem. As we approach the development of advanced artificial intelligence, the structure of value and our ability to specify it has become perhaps the most critical challenge facing humanity. This challenge assumes…

Capability Control vs Motivation Selection: Contrasting Strategies for AI Safety
|

Capability Control vs Motivation Selection: Contrasting Strategies for AI Safety

Capability Control: The Tool or Slave Model Capability control focuses on limiting what an AI can do, rather than shaping why it does it. The idea is to ensure AI systems behave predictably and stay within strict bounds – like a powerful but obedient tool. This includes: In this model, the AI is treated less…

VRisk – Value Risk
|

VRisk – Value Risk

This is not an attempt at repackaging known concepts, it’s an attempt to carve out axis of concern, hopefully to help increase clarity and weight to concerns about value choice – what would otherwise sound like a vague fear of “wrong values.” What is Value Risk? Value Risk (VRisk) is the risk that a system—technological…