I'm worried your subdivision misses a significant proportion of harms that don't fall into either category. For instance, interactions that don't involve malice or power dynamics and are innocuous in isolation but harmful when repeated. This repetition can be made more likely by imbalanced gender ratios.
I think being flirted with during the day at an EAG like Nathan discussed above is a good example of this. If you're flirted with once over the weekend, perhaps it's fine or even nice, especially if it's from the person you found most interesting. But if you're flirted with several times, you may start to feel uncomfortable.
Well if a conference has 3x more men than woman and 1-on-1s are matched uniformly at random, then women have 3x more cross-gender 1-on-1s than men. Assuming all people are equally likely to flirt with someone of a different gender than them, it's very possible that the average man receives a comfortable amount of flirting while the average woman receives an uncomfortable amount.
And it probably gets worse when one considers that these are random variables and we don't care about the average but rather about how many people exceed the uncomfortable threshold and to what degree. And perhaps worse again if certain "attractive" people are more likely to receive flirting.
Overall, my point is that behaviors and norms that would be fine with balanced gender ratios can be harmful with imbalanced ones. Unfortunately, we have imbalanced ones and we need to adapt accordingly.
Also, the quoted passage seems to assume that EA orgs optimize for their org’s impact rather than for the impact of the movement/good of the world. I’m not convinced that’s true. I would be surprised if EA orgs were attempting to poach workers they explicitly believed were having more impact at other organizations.
It does seem possible that orgs overestimate their own impact/the impact of roles they hire for. However, this would still lead to a much smaller effect than if they completely ignore the impact of candidates at their current roles, as the post seems to assume.
Thanks for link-posting, I enjoyed this!
I didn't understand the section about EA being too centralized and focused on absolute advantage. Can anyone explain?
EA-in-practice is too centralized, too focused on absolute advantage; the market often does a far better job of providing certain kinds of private (or privatizable) good. However, EA-in-practice likely does a better job of providing certain kinds of public good than do many existing institutions.
And footnote 11:
It's interesting to conceive of EA principally as a means of providing public goods which are undersupplied by the market. A slightly deeper critique here is that the market provides a very powerful set of signals which aggregate decentralized knowledge, and help people act on their comparative advantage. EA, by comparison, is relatively centralized, and focused on absolute advantage. That tends to centralize people's actions, and compounds mistakes. It's also likely a far weaker resource allocation model, though it does have the advantage of focusing on public goods. I've sometimes wondered about a kind of "libertarian EA", more market-focused, but systematically correcting for well-known failures of the market.↩
Don't global health charities provide private goods (bed nets, medicine) that markets cannot? Markets only supply things people will pay for and poor people can't pay much.
X-risk reduction seems like a public good, and animal welfare improvements are either a public good of a private good where the consumers definitely cannot pay.
I take it that centralization is in contrast to markets. But it seems like in a very real way EA is harnessing markets to provide these things. EA-aligned charities are competing in a market to provide QALYs as cheaply as possible, since EAs will pay for them. EAs also seem very fond of markets generally (ex: impact certificates, prediction markets).
How is EA focused on absolute advantage? Isn't earning to give using one's relative advantage?
Although it's against longtermism and not EA, this recent blog post by Boaz Barak (a computer science professor at Harvard) might qualify.
Great post, James!
I'm curious if you have any sense of how the average conditions/welfare levels of farmed animals are expected to change on this default trajectory, or how they've changed in the last few decades. I imagine this is difficult to quantify, but seems important.
In particular, assuming market pressures stay as they are, how should we expect technological improvements to affect farmed animal welfare?
My uneducated guess: optimizing hard for (meat production / cost) generally leads to lower animal welfare. This seems roughly true of technological improvements in the past. For example: