(This post also has a Russian version, translated from the present original by K. Kirdan.)
Many longtermists seem hopeful that our successors (or any advanced civilization/superintelligence) will eventually act in accordance with some moral truth.[1] While I’m sympathetic to some forms of moral realism, I believe that such a scenario is fairly unlikely for any civilization and even more so for the most advanced/expansionist ones. This post briefly explains why.
To be clear, my case does under no circumstances imply that we should not act according to what we think might be a moral truth. I simply argue that we can't assume that our successors -- or any powerful civilization -- will "do the (objectively) right thing". And this matters for longtermist cause prioritization.
Epistemic status: Since I believe the ideas in this post to be less important than those in future ones within this sequence, I wrote it quickly and didn’t ask anyone for thorough feedback before posting, which makes me think I’m more likely than usual to have missed important considerations. Let me know what you think!
Update April 10th: When I first posted this, the title was "It Doesn't Matter what the Moral Truth might be". I realized this was misleading. It was making it look like I was making a strong normative claim regarding what matters while my goal was actually to predict what might happen, so I changed it.
Rare are those who will eventually act in accordance with some moral truth
For agents to do what might objectively be the best thing to do, you need all these conditions to be met:
- There is a moral truth.
- It is possible to “find it” and recognize it as such.
- They find something they recognize as a moral truth.
- They (unconditionally) accept it, even if it is highly counterintuitive.
- The thing they found is actually the moral truth. No normative mistake.
- They succeed at acting in accordance with it. No practical mistake.
- They stick to this forever. No value drift.
I think these seven conditions are generally quite unlikely to be all met at the same time, mainly for the following reasons:
- (Re: condition #1) While I find compelling the argument that (some of) our subjective experiences are instantiations of objective (dis)value (see Rawlette 2016; Vinding 2014), I am highly skeptical about claims of moral truths that are not completely dependent on sentience.
- (Re: #2) I don’t see why we should assume it is possible to “find” (with a sufficient degree of certainty) the moral truth, especially if it is more complex than – or different from – something like “pleasure is good and suffering is bad.”
- (Re: #3 and #4) If they “find” a moral truth and don’t like what it says, they might very well not act in accordance with it?[2]
- (Re: #3, #4, #5, and #7) Within a civilization, we should expect the agents who have the values that are the most adapted/competitive to survival, replication, and expansion to eventually be selected for (see, e.g., Bostrom 2004; Hanson 1998), and I see no reason to suppose the moral truth is particularly well adapted to those things.
Even if they’re not rare, their impact will stay marginal
Now, let’s actually assume that many smart agents converge on THE moral truth and effectively optimize for whatever it says. The thing is that, for reasons analogous to what we addressed in the last bullet point above, we may expect civilizations -- or groups/individuals within a civilization -- adopting the moral truth to be less competitive than those who have the values that are the most adaptive and adapted to space colonization races.
My subsequent post investigates this selection effect in more detail, but here is an intuition pump: Say Denmark wants to follow the moral truth which is to maximize the sum , where is something valuable and something disvaluable. Meanwhile, France just wants something close to “occupy as much space territory as possible”. While the Danes face a trade-off between (A) spreading and building military weapons/defenses as fast as possible and (B) investing in “colonization safety”[3] to make sure they actually end up optimizing for what the moral truth says, the French don't face this trade-off and can just go all-in on (A), which gives them an evolutionary advantage. The significance of this selection effect, here, depends on whether the moral truth is among – or close to – the most “expansion-conducive” intrinsic goals civilizations can plausibly have, and I doubt that it is.
Conclusion
Acting in accordance with some moral truth requires the unlikely successful succession of many non-obvious steps.
Also, values don’t come out of nowhere. They are the product of evolutionary processes. We should expect the most adaptive and adapted values to be the most represented, at least among the most expansionist societies. And how true a moral theory might be seems fairly orthogonal to how competitive it is,[4] such that we -- a priori -- have no good reason to expect (the most powerful) civilizations/agents to do what might be objectively good.
If I'm roughly correct, this implies that the "discoverable moral reality" argument in favor of assuming the future will be good (see Anthis 2022) is pretty bad. This probably also has more direct implications for longtermist cause prioritization that will be addressed in subsequent posts within this sequence.
Acknowledgment
My work on this sequence so far has been funded by Existential Risk Alliance.
All assumptions/claims/omissions are my own.
- ^
This is informed by informal interactions I had, plus my recollections of claims made in some podcasts I can’t recall. I actually can’t find anything fleshing out this exact idea, surprisingly, and I don’t think it’s worth spending more time searching. Please, share in the comments if you can think of any!
- ^
Interestingly, Brian Tomasik (2014) writes: “Personally, I don't much care what the moral truth is even if it exists. If the moral truth were published in a book, I'd read the book out of interest, but I wouldn't feel obligated to follow its commands. I would instead continue to do what I am most emotionally moved to do.”
- ^
After a thorough evaluation, they might even realize that the best way to maximize their utility requires avoiding colonizing space (e.g., because the expected disvalue of conflict with France or conflict with alien civilizations is too high).
- ^
This comment thread discusses an interesting argument of Wei Dai that challenges this claim.
In Six Plausible Meta-Ethical Alternatives, I wrote (as one of the six alternatives):
I think in this post you're not giving enough attention to the possibility that there's something that we call "doing philosophy" that can be used to discover all kinds of philosophical truths, and that you can't become a truly powerful civilization without being able to "do philosophy" and be generally motivated by the results. Consider that philosophy seems to have helped the West become the dominant civilization on Earth, for example by inventing logic and science, and more recently have led to the discovery of ideas like acausal extortion/trade (which seem promising albeit still highly speculative). Of course I'm very uncertain of this and have little idea what "doing philosophy" actually consists of, but I've written a few more words on this topic if you're interested.
Very interesting, Wei! Thanks a lot for the comment and the links.
TL;DR of my response: Your argument assumes that the first two conditions I list are met by default, which is I think a strong assumption (Part 1). Assuming that is the case, however, your point suggests there might be a selection effect favoring agents that act in accordance with the moral truth, which might be stronger than the selection effect I depict for values that are more expansion-conducive than the moral truth. This is something I haven't seriously considered and this made me update! Nonetheless, for your argument to be valid and strong, the orthogonality thesis has to be almost completely false, and I think we need more solid evidence to challenge that thesis (Part 2).
Part 1: Strong assumption
My understanding is that this scenario says the seven conditions I listed are met because it is actually trivial for a super-capable intergalactic civilization to meet those (or even required for it to become intergalactic in the first place, as you suggest later).
I think this is plausible for the following conditions:
You might indeed expect that the most powerful civs figure out how to overcome these challenges, and that those who don't are left behind.[1] This is something I haven't seriously considered before, so thanks!
However, recall the first two conditions:
How capable a civilization is doesn't matter when it comes to how likely these two are to be met. And while most metaethical debates focus only on 1, saying 1 is true is a much weaker claim than saying 1&2 is true (see, e.g., the naturalism vs non-naturalism controversy, which is I think only one piece of the puzzle).
Part 2: Challenging the orthogonality thesis
Then, you say that in this scenario you depict
Maybe, but what I argue I that they are (occasional) "sophisticated minds" with values that are more expansion-conducive than the (potential) moral truth (e.g., because they have simple unconstrained goals such as "let's just maximize for more life" or "for expansion itself"), and that they're the ones who tend to take over.
But then you make this claim, which, if true, seems to sort of debunk my argument:
(Given the context in your comment, I assume that by "being able to do philosophy", you mean "being able to do things like finding the moral truth".)
But I don't think this claim is true.[1] However, you made me update and I might update more once I read the posts of yours that you linked! :)
I remain skeptical because this would imply the orthogonality thesis is almost completely false. Assuming there is a moral truth and that it is possible to "find" it and recognize it as such, I tentatively still believe that extremely powerful agents/civs with motivations misaligned with the moral truth are very plausible and not rare. You can at least imagine scenarios where they started aligned but then value drifted (without that making them significantly less powerful).
The orthogonality thesis could be (and I think almost certainly is) false with respect to some agent-generating processes (e.g., natural selection) and true with respect to others (e.g. Q-learning).
Do you have any reading to suggest on that topic? I'd be curious to understand that position more :)
My impression was that philosophers tended to disagree a lot on what moral truths are?
I'd argue that the process of Western civilization dominating the Earth was not a very moral process, and was actually pretty immoral, despite the presence of logic and science. It involved several genocides (in the Americas), colonization, the Two World Wars... In the process, some good things definitely happened (medicine progressing, for instance), but mostly for humans. The status of farmed animals seems to have consistently worsened with factory farming stepping in.
So I'd argue that was the West dominating the world happened because it was more powerful, not because it was more moral (see The End of the Magemachine by Fabian Scheidler for more on this topic).
In that view, science and logic matter because they allow you to have more power. They allow to have a more truthful picture of how the universe works, which allows making stuff like firearms and better boats and antibiotics and nuclear bombs. But this is the process of "civilizations competing with each other" described above. It's not a comparison based on "who is acting closer to what is morally good"?
Several objections in no particular order
> Within a civilization, we should expect the agents who have the values that are the most adapted/competitive to survival, replication, and expansion to eventually be selected for
is the right conclusion to draw from Hanson's paper. The selection effects he's talking about act on different regions of the frontier of a given civilization. Those living in the interior (who will be the vast majority of the population in the extreme long-run) may be disproportionately descended from fast expanders, but will not face the same pressure themselves.
Insightful! Thanks for taking the time to write these.
Agreed and I didn't mean to argue against that so thanks for clarifying! Note however that the more you expect the moral truth to be fragile/complex, the further from it you should expect agents' actions to be.
Hum... I don't think the "such that..." part logically follows. I don't think this is how selection effects work. All I'm saying is that those who are the most bullish on space colonization will colonize more space.
I'm not sure what to say regarding your last two points. I think I need to think/read more, here. Thanks :)
Sure, but that doesn't tell you much about what happens afterwards. If the initial colonists' values are locked in ~forever, we should probably expect value drift to be weak in general, which means frontier selection effects have a lot less variation to work with.
At the extreme lower limit with no drift at all, most agents within a mature civilization are about as expansionist as the most expansionist of the initial colonists - but no more so. And this might not be all that much in the grand scheme of things.
At the other end, where most of the space of possible values gets explored, maybe you do get a shockwave of superintelligent sociopaths racing outwards at relativistic speeds - but you also get a vast interior that favors (relatively speaking) long-term survival and material efficiency.
Hi Jim,
Nice post!
Under hedonism, any and all claims of moral truths are completely dependent on sentience. So 100 % weight on hedonism (in principle) makes condition 1 met, right?
In general, the greater the amount of resources of a civilisation, the more resources could be directed towards finding the moral truth. So should we expect more powerful civilisations to be better at finding the moral truth? I suppose one could argue that such civilisations may be selected such that they direct fewer resources towards finding the moral truth, but I do not think that is obvious, because finding the moral truth also has selection effects. For example, finding the moral truth seIects for not going extinct (to continue searching for it), and space colonisation arguably decreases extinction risk.