T

Taymon

424 karmaJoined

Posts
1

Sorted by New

Comments
55

A number of people invited me to 1:1s to ask me for career advice in my field, which is software engineering. Mostly of the "how do I get hired" kind rather than the "how do I pick a career path that's most in line with EA strategic priorities" kind that 80,000 Hours specializes in. Unfortunately I'm not very good at this kind of advice (I haven't looked for a new job in more than eight years) and haven't been able to find anywhere else I could send people to that would be more helpful. I think there used to be an affinity group or something for EA software engineers, but I don't think it's active anymore.

Anyone know of anything like this? If not, and if you're the kind of person who's well-positioned to start a group like this, consider this a request for one.

Honestly it seems kind of weird that on the EA Forum there isn't just a checkbox for this.

I've often thought that there should be separate "phatic" and "substantive" comment sections.

The Fun Theory Sequence (which is on a similar topic) had some things to say about the Culture.

In the last paragraph, did you mean to write "the uncertainty surrounding the expected value of each policy option is high"?

While true, I think most proposed EA policy projects are much too small in scope to be able to move the needle on trust, and so need to take the currently-existing level of trust as a given.

I agree that that the word ‘populism’ is very prone to misunderstandings but I think the term 'technocracy' is acceptably precise. While precision is important, I think we should balance this against the benefits of using more common words, which make it easier for the reader to make connections with other arguments in favour of or against a concept.

I should clarify: I think the misunderstandings are symptoms of a deeper problem, which is that the concept of "technocracy" is too many different things rolled into one word. This isn't about jargon vs. non-jargon; substituting a more jargon-y word doesn't help. (I think this is part of why it's taken on such negative connotations, because people can easily roll anything they don't like into it; that's not itself a strong reason not to use it, but it's illustrative.)

"Technocracy" works okay-ish in contexts like this thread where we're all mostly speaking in vague generalities to begin with, but when discussing specific policies or even principles for thinking about policy, "I think this is too technocratic" just isn't helpful. More specific things like "I think this policy exposes the people executing it to too much moral hazard", or "I think this policy is too likely to have unknown-unknowns that some other group of people could have warned us about", are better. Indeed, those are very different concerns and I see no reason to believe that EA-in-general errs the same amount, or even in the same direction, for each of them. (If words like "moral hazard" are too jargon-y then you can just replace them with their plain-English definitions.)

I also think that EAs haven't sufficiently considered populism as a tool to deal with moral uncertainty.

I agree that there hasn't been much systematic study of this question (at least not that I'm aware of), and maybe there should be. That being said, I'm deeply skeptical that it's a good idea, and I think most other EAs who've considered it are too, which is why you don't hear it proposed very often.

Some reasons for this include:

  • The public routinely endorses policies or principles that are nonsensical or would obviously result in terrible outcomes. Examples include Philip Tetlock's research on taboo tradeoffs [PDF], and this poll from Reuters (h/t Matt Yglesias): "Nearly 70 percent of Americans, including a majority of Republicans, want the United States to take 'aggressive' action to combat climate change—but only a third would support an extra tax of $100 a year to help."
  • You kind of can't ask the public what they think about complicated questions; they're very diverse and there's a lot of inferential distance. You can do things like polls, but they're often only proxies for what you really want to know, and pollster degrees-of-freedom can cause the results to be biased.
  • When EAs look back on history, and ask ourselves what we would/should have done if we'd been around then—particularly on questions (like whether slavery is good or bad) whose morally correct answers are no longer disputed—it seems to look like we would/should have sided with technocrats over populists, much more often than the reverse. A commonly-cited example is William Wilberforce, largely responsible for the abolition of slavery in the British Empire. Admittedly, I'd like to see some attempt to check how representative this is (though I don't expect that question to be answerable comprehensively).
Load more