Davidmanheim

Head of Research and Policy @ ALTER - Association for Long Term Existence and Resilience
7043 karmaJoined Working (6-15 years)

Participation
4

  • Received career coaching from 80,000 Hours
  • Attended more than three meetings with a local EA group
  • Completed the AGI Safety Fundamentals Virtual Program
  • Completed the In-Depth EA Virtual Program

Sequences
2

Deconfusion and Disentangling EA
Policy and International Relations Primer

Comments
821

Topic contributions
1

"Neither of which current LLMs appear to be capable of."

If o1 pro isn't able to both hack and get money yet, it's shockingly close. (Instruction tuning for safety makes accessing that capability very difficult.)

My tentative take is that this is on-net bad, and should not be encouraged. I give this a 10/10 for good intent, but a 2/10 for planning and avoiding foreseeable issues, including the unilateralists curse, the likely object level impacts of the pledge, and the reputational and community impacts of promoting the idea.

It is not psychologically healthy to optimize or maximize your life towards a single goal, much less commit to doing so. That isn't the EA ideal. Promising to "maximize my ability to make a meaningful difference" is an unlimited and worryingly cult-like commitment, builds in no feedback from others who have a broader perspective about what is or is not important or useful. It implicitly requires pledgers to prioritize impact over personal health and psychological wellbeing. (The claim that it's usually the case that burnout reduces impact is a contingent one, and seems very likely to lead many people to overcommit and do damaging things.) It leads to unhealthy competitive dynamics, and excludes most people, especially the psychologically well adjusted.

I will contrast this to the giving pledge, which is very explicitly a partial pledge, requiring 10% of your income. This is achievable without extreme measures, or giving up having a normal life. The pledge was built via consultation with and advice from a variety of individuals, especially including those who were more experienced, which also seems to sharply contrast with this one.

I've referred to this latter point as candy bar extinction; using fixed discount rates, a candy bar is better than preventing extinction with certainty after some number of years. (And with moderately high discount rates, the number of years isn't even absurdly high!)

Thanks - this is helpful as a term, and closely related to privileging the hypothesis; https://www.lesswrong.com/posts/X2AD2LgtKgkRNPj2a/privileging-the-hypothesis The general solution, of course,is expensive but necessary; https://secondenumerations.blogspot.com/2017/03/episode-6-method-of-multiple-working.html

Worth noting that a number of 1DaySooner research projects I worked on or ran have paid some undergraduates, grad students, and medical school students for supervised work on research projects, which is effectively very similar to a paid internship - but as you mentioned, it's very hard to do so outside a well scoped project.

I've written about this here, where I said, among other things:

Obviously, charity is a deeply personal decision - but it’s also a key way to impact the world, and an expression of religious belief, and both are important to me. Partly due to my experience, I think it’s important to dedicate money to giving thoughtfully and in advance, rather than doing so on an ad-hoc basis - and I have done this since before hearing about Effective Altruism. But inspired by Effective Altruism and organizations like Givewell, I now dedicate 10% of my income to charities that have been evaluated for effectiveness, and which are aligned with my beliefs about charitable giving.

In contrast to the norm in effective altruism, I only partially embrace cause neutrality. I think it’s an incomplete expression of how my charity should impact the world. For that reason, I split my charitable giving between effective charities which I personally view as valuable, and deference to cause-neutral experts on the most impactful opportunities. Everyone needs to find their own balance, and I have tremendous respect for people who donate more, but I’ve been happy with my decision to limit my effective charitable giving at 10%, and beyond that, I still feel free to donate to other causes, including those that can’t be classified as effective at all.

As suggested above, community is an important part of my budget. A conclusion I came to after reflecting on the question, and grappling with effective altruism, is that separate from charitable giving, I think it’s important to pay for public goods you benefit from, both narrow ones like community organizations, and broader ones. I think it’s worth helping to fund community centers, and why I paid for NPR membership when I lived in the US, and why I pay to offset carbon emissions to reduce the harms of climate change
 

This is the wrong thing to try to figure out; most of the probability of existential risk is  likely not to make a clear or intelligible story. Quoting Nick Bostrom:
 

Suppose our intuitions about which future scenarios are “plausible and realistic” are shaped by what we see on TV and in movies and what we read in novels. (After all, a large part of the discourse about the future that people encounter is in the form of fiction and other recreational contexts.) We should then, when thinking critically, suspect our intuitions of being biased in the direction of overestimating the probability of those scenarios that make for a good story, since such scenarios will seem much more familiar and more “real”. This Good-story bias could be quite powerful. When was the last time you saw a movie about humankind suddenly going extinct (without warning and without being replaced by some other civilization)? While this scenario may be much more probable than a scenario in which human heroes successfully repel an invasion of monsters or robot warriors, it wouldn’t be much fun to watch. So we don’t see many stories of that kind. If we are not careful, we can be mislead into believing that the boring scenario is too farfetched to be worth taking seriously. In general, if we think there is a Good-story bias, we may upon reflection want to increase our credence in boring hypotheses and decrease our credence in interesting, dramatic hypotheses. The net effect would be to redistribute probability among existential risks in favor of those that seem to harder to fit into a selling narrative, and possibly to increase the probability of the existential risks as a group.

If you're "crossposting," you should at least copy and paste the intro section, to give people something to look at other than the title:

It would be bad if your city burned down.  This is a reasonable fear to have.  What to do, then, if your neighbor’s house is on fire?  Surely not blame the fire department for the fire and organize a protest in front of the station to trap the firetrucks inside.  That's no longer fear, but self-destructive panic.  Hysteria makes the city burn down.

It would be bad if there were a third world war.  That is a reasonable fear to have.  By resisting Russia, Ukrainians are making every scenario for such a catastrophe less likely.  On the scale of our world, the Ukrainians are the firemen.  They are keeping the rest of us safe.  It makes no sense to blame them for Russia's invasion, nor to hinder them from doing their job.  That's no longer fear, but self-destructive panic.  Hysteria makes the third world war more likely.

Let's consider, soberly, three familiar scenarios for a third world war: (1) escalation from a conventional war in Europe; (2) escalation from a conventional war in the Pacific; and (3) the spread of nuclear weapons.  In all three cases, Ukrainian resistance makes the rest of us safer.  The Ukrainians are containing the ongoing war in Europe to their own country; they are deterring the war in the Pacific; and they are preventing the spread of nuclear weapons.  These scenarios for a third world war have not played out because Ukrainians take risks.

...

As an aside, the reports end up being critical in worlds where the rapid response is needed, since they show ongoing attention, and will be looked at in retrospect. But they can also be used more directly to galvanize news coverage on key topics and as evidence by policy orgs. Promoting that avenue for impact seems valuable.

Responding to only one minor point you made, the 6-month pause letter seems like the type of thing you oppose: it's not able to help with the risk, it just does deceptive PR that aligns with the goals of pushing against AI progress, while getting support from those who disagree with their actual goal.

Load more