RN

richard_ngo

7508 karmaJoined

Bio

Former AI safety research engineer, now AI governance researcher at OpenAI. Blog: thinkingcomplete.blogspot.com

Sequences
2

Replacing Fear
EA Archives Reading List

Comments
331

Thanks for sharing this, it does seem good to have transparency into this stuff.

My gut reaction was "huh, I'm surprised about how large a proportion of these people (maybe 30-50%, depending on how you count it) I don't recall substantially interacting with" (where by "interaction" I include reading their writings).

To be clear, I'm not trying to imply that it should be higher; that any particular mistakes are being made; or that these people should have interacted with me. It just felt surprising (given how long I've been floating around EA) and worth noting as a datapoint. (Though one reason to take this with a grain of salt is that I do forget names and faces pretty easily.)

My point is not that the current EA forum would censor topics that were actually important early EA conversations, because EAs have now been selected for being willing to discuss those topics. My point is that the current forum might censor topics that would be important course-corrections, just as if the rest of society had been moderating early EA conversations, those conversations might have lost important contributions like impartiality between species (controversial: you're saying human lives don't matter very much!), the ineffectiveness of development aid (controversial: you're attacking powerful organizations!), transhumanism (controversial, according to the people who say it's basically eugenics), etc.

Re "conversations can be had in more sensitive ways", I mostly disagree, because of the considerations laid out here: the people who are good at discussing topics sensitively are mostly not the ones who are good at coming up with important novel ideas.

For example, it seems plausible that genetic engineering for human intelligence enhancement is an important and highly neglected intervention. But you had to be pretty disagreeable to bring it into the public conversation a few years ago (I think it's now a bit more mainstream).

Narrowing in even further on the example you gave, as an illustration: I just had an uncomfortable conversation about age of consent laws literally yesterday with an old friend of mine. Specifically, my friend was advocating that the most important driver of crime is poverty, and I was arguing that it's cultural acceptance of crime. I pointed to age of consent laws varying widely across different countries as evidence that there are some cultures which accept behavior that most westerners think of as deeply immoral (and indeed criminal).

Picturing some responses you might give to this:

  1. That's not the sort of uncomfortable claim you're worried about
    1. But many possible continuations of this conversation would in fact have gotten into more controversial territory. E.g. maybe a cultural relativist would defend those other countries having lower age of consent laws. I find cultural relativism kinda crazy (for this and related reasons) but it's a pretty mainstream position.
  2. I could have made the point in more sensitive ways
    1. Maybe? But the whole point of the conversation was about ways in which some cultures are better than others. This is inherently going to be a sensitive claim, and it's hard to think of examples that are compelling without being controversial.
  3. This is not the sort of thing people should be discussing on the forum
    1. But EA as a movement is interested in things like:

      1. Criminal justice reform (which OpenPhil has spent many tens of millions of dollars on)
      2. Promoting women's rights (especially in the context of global health and extreme poverty reduction)
      3. What factors make what types of foreign aid more or less effective
      4. More generally, the relationship between the developed and the developing world

      So this sort of debate does seem pretty relevant.

I think EA would've broadly survived intact by lightly moderating other kinds of discomfort (or it may have even expanded).

The important point is that we didn't know in advance which kinds of discomfort were of crucial importance. The relevant baseline here is not early EAs moderating ourselves, it's something like "the rest of academic philosophy/society at large moderating EA", which seems much more likely to have stifled early EA's ability to identify important issues and interventions.

(I also think we've ended up at some of the wrong points on some of these issues, but that's a longer debate.)

Ty for the reply; a jumble of responses below.

I think there are better places to have these often awkward, fraught conversations.

You are literally talking about the sort of conversations that created EA. If people don't have these conversations on the forum (the single best way to create common knowledge in the EA commmunity), then it will be much harder to course-correct places where fundamental ideas are mistaken. I think your comment proceeds from the implicit assumption that we're broadly right about stuff, and mostly just need to keep our heads down and do the work. I personally think that a version of EA that doesn't have the ability to course-correct in big ways would be net negative for the world. In general it is not possible to e.g. identify ongoing moral catastrophes when you're optimizing your main venue of conversations for avoiding seeming weird.

I agree with you the quote from the Hamas charter is more dangerous - and think we shouldn't be publishing or discussing that on the forum either.

If you're not able to talk about evil people and their ideologies, then you will not be able to account for them in reasoning about how to steer the world. I think EA is already far too naive about how power dynamics work at large scales, given how much influence we're wielding; this makes it worse.

There's potential reputational damage for all the people doing great EA work across the spectrum here.

I think there are just a few particular topics which give people more ammunition for public take-downs, and there is wisdom in sometimes avoiding loading balls into your opponents cannons.

Insofar as you're thinking about this as a question of coalitional politics, I can phrase it in those terms too: the more censorious EA becomes, the more truth-seeking people will disaffiliate from it. Habryka, who was one of the most truth-seeking people involved in EA, has already done so; I wouldn't say it was directly because of EA not being truth-seeking enough, but I think that was one big issue for him amongst a cluster of related issues. I don't currently plan to, but I've considered the possibility, and the quality of EA's epistemic norms is one of my major considerations (of course, the forum's norms are only a small part of that).

However, having said this, I don't think you should support more open forum norms mostly as a concession to people like me, but rather in order to pursue your own goals more effectively. Movements that aren't able to challenge foundational assumptions end up like environmentalists: actively harming the causes they're trying to support.

I appreciate the thought that went into this. I also think that using rate-limits as a tool, instead of bans, is in general a good idea. I continue to strongly disagree with the decisions on a few points:

  1. I still think including the "materials that may be easily perceived as such" clause has a chilling effect.
  2. I also remember someone's comment that the things you're calling "norms" are actually rules, and it's a little disingenuous to not call them that; I continue to agree with this.
  3. The fact that you're not even willing to quote the parts of the post that were objectionable feels like an indication of a mindset that I really disagree with. It's like... treating words as inherently dangerous? Not thinking at all about the use-mention distinction? I mean, here's a quote from the Hamas charter: "There is no solution for the Palestinian question except through Jihad." Clearly this is way way more of an incitement to violence than any quote of dstudiocode's, which you're apparently not willing to quote. (I am deliberately not expressing any opinion about whether the Hamas quote is correct; I'm just quoting them.) What's the difference?
  4. "They see the fact that it is “just” a philosophical question as not changing the assessment." Okay, let me now quote Singer. "Human babies are not born self-aware, or capable of grasping that they exist over time. They are not persons... the life of a newborn is of less value than the life of a pig, a dog, or a chimpanzee.” Will you warn/ban me from the EA forum for quoting Singer, without endorsing that statement? What if I asked, philosophically, "If Singer were right, would it be morally acceptable to kill a baby to save a dog's life?" I mean, there are whole subfields of ethics based on asking about who you would kill in order to save whom (which is why I'm pushing on this so strongly: the thing you are banning from the forum is one of the key ways people have had philosophical debates over foundational EA ideas). What if I defended Singer's argument in a post of my own?

As I say this, I feel some kind of twinge of concern that people will find this and use it to attack me, or that crazy people will act badly inspired by my questions. I hypothesize that the moderators are feeling this kind of twinge more generally. I think this is the sort of twinge that should and must be overridden, because listening to it means that your discourse will forever be at the mercy of whoever is most hostile to you, or whoever is craziest. You can't figure out true things in that situation.

(On a personal level, I apologize to the moderators for putting them in difficult situations by saying things that are deliberately in the grey areas of their moderation policy. Nevertheless I think it's important enough that I will continue doing this. EA is not just a group of nerds on the internet any more, it's a force that shapes the world in a bunch of ways, and so it is crucial that we don't echo-chamber ourselves into doing crazy stuff (including, or especially, when the crazy stuff matches mainstream consensus). If you would like to warn/ban me, then I would harbor no personal ill-will about it, though of course I will consider that evidence that I and others should be much more wary about the quality of discourse on the forum.)

This moderation policy seems absurd. The post in question was clearly asking purely hypothetical questions, and wasn't even advocating for any particular answer to the question. May as well ban users for asking whether it's moral to push a man off a bridge to stop a trolley, or ban Peter Singer for his thought experiments about infanticide.

Perhaps dstudiocode has misbehaved in other ways, but this announcement focuses on something that should be clearly within the bounds of acceptable discourse. (In particular, the standard of "content that could be interpreted as X" is a very censorious one, since you now need to cater to a wide range of possible interpretations.)

Load more