Bias in the Extrapolation Base: The Silent War Over AI’s Values
|

Bias in the Extrapolation Base: The Silent War Over AI’s Values

Nick Bostrom addresses concerns about biased influence on the extrapolation base in his discussions on indirect normativity (IN – i.e., an approach where the AI deduces ideal values or states of affairs rather than running with current values and states of affairs) esp. coherent extrapolated volition (CEV) and value alignment in his magnum opus Superintelligence….

The Is and the Ought

The Is and the Ought

Back to basics. The “is” is what science and rationality tell us about the world and logic – and “ought” represents moral obligations, values, or prescriptions about how the world should be, rather than how it is. What is an “is”? The “is” refers to factual statements about the world, encompassing empirical observations, logical truths…

Survival, Cooperation, and the Evolution of Values in the Age of AI
|

Survival, Cooperation, and the Evolution of Values in the Age of AI

No one in their right mind can deny the struggle to survive animates natural selection. Though it’s a bit misleading to say that ‘survival is paramount – and that every strategy serves that goal’: survival plays a crucial role in evolution, but it is not the ultimate goal. As emphasised by Charles Darwin’s theory of…

Human Values Approximate Ideals in Objective Value Space

Human Values Approximate Ideals in Objective Value Space

Value Space and “Good” Human Values Human values can be conceptualised as occupying regions in a vast objective multidimensional “value space.” These regions reflect preferences for cooperation, survival, flourishing, and minimising harm, among other positive traits. If TAI / SI can approximate the subset of human values deemed “good” (e.g., compassion, fairness, cooperation), while avoiding…

Early Philosophical Groundwork for Indirect Normativity
|

Early Philosophical Groundwork for Indirect Normativity

Note: this is by no means exhaustive. it’s focus is on western philosophy. I’ll expand and categorise more later… The earliest precursors to indirect normativity can be traced back to early philosophical discussions on how to ground moral decision-making in processes or frameworks rather than specific, static directives. While Nick Bostrom’s work on indirect normativity…

Kristian Rönn – The Darwinian Trap – Interview with SciFuture
|

Kristian Rönn – The Darwinian Trap – Interview with SciFuture

Kristian Rönn discusses his amazing book The Darwinian Trap – Macroeconomics & Game theoretic Imagineering: avoiding the minefields of the tragedy of the common in the hope that we can stave off the onslaught of Darwinian Demons, and ultimately find some nash equilibrium of fairness for all humanity, sentient AI an non-human animals. The book…

Understanding V-Risk: Navigating the Complex Landscape of Value in AI
|

Understanding V-Risk: Navigating the Complex Landscape of Value in AI

Will AI Preserve Our Values, or Improve on them? Imagine a future where advanced AIs follow every instruction flawlessly – but humanity feels strangely adrift. Our tools are obedient, but the ‘soul’ of our civilisation feels absent. This is the hidden danger of V-risk—value erosion. In this post I explore what I explore what I…

Value Space
|

Value Space

The concept of a non-arbitrary objective value space offers a compelling framework for understanding how to navigate to more valuable worlds. This framework presupposes the existence of stance-independent moral truths – ethical principles that rational agents will converge upon given sufficient cognitive sophistication [1]. Human values are a narrow slice of value space – In…