Does anyone know of a low-hassle way to charge invoices for services but it's a third-party charity that gets paid? It could well be an EA charity if that makes it easy. I'm hoping for something slightly more structured than "I'm not receiving any pay for my services but I'm trusting you to donate X amount to this charity instead".
I've seen EA writing (particularly about AI safety) that goes something like:
I know X and Y thought leaders in AI safety, they're exceptionally smart people with opinion A, so even though I personally think opinion B is more defensible, I also think I should be updating my natural independent opinion in the direction of A, because they're way smarter and more knowledgeable than me.
I'm struggling to see how this update strategy makes sense. It seems to have merit when X and Y know/understand things that literally no other expert knows, but aside from that, in all other scenarios that come to mind, it seems neutral at best, otherwise a worse strategy than totally disregarding the "thought leader status" of X and Y.
Am I missing something?
The reasoning is that knowledgeable people's beliefs in a certain view is evidence for that view.
This is a type of reasoning people use a lot in many different contexts. I think it's a valid and important type of reasoning (even though specific instances of it can of course be mistaken).
Some references:
https://plato.stanford.edu/entries/disagreement/#EquaWeigView
https://www.routledge.com/Why-Its-OK-Not-to-Think-for-Yourself/Matheson/p/book/9781032438252
https://forum.effectivealtruism.org/posts/WKPd79PESRGZHQ5GY/in-defence-of-epistemic-modesty
What you describe in your first paragraph sounds to me like a good updating strategy, except I would say that you’re not updating your “natural independent opinion,” you’re updating your all-things-considered belief.
Related short posts I recommend—the first explains the distinction I’m pointing at, and the second shows how things can go wrong if people don’t track it:
Isn't credentialism basically a form of ad hominem attack?
My take on this is inspired by the Fallacy Fork, which is to say that there are two ways in which we can understand "ad hominem" (among many other fallacies):
There are still some reasons to avoid credentialism or ad hominem for reasons other than maximising the truth of the claim you're considering:
The point of credentialism is that the ideal circumstances for an individual to evaluate ideas don't exist very often. Medical practitioners aren't always right, and homeopaths or opinion bloggers aren't always wrong, but bearing in mind that I'm seldom well enough versed in the background literature to make my own mind up, trusting the person with the solid credentials over the person with zero or quack credentials is likely to be the best heuristic in the absence of any solid information to the contrary
And yes, of course sometimes it isn't, and sometimes the bar is completely arbitrary (the successful applicant will have some sort of degree from some sort of top 20 university) or the level of distinction irrelevant (his alma mater is more reputable than hers) and sometimes the credentials themselves are suspect
One of the canonical EA books (can't remember which) suggests that if an individual stops consuming eggs (for example), almost all the time this will have zero impact, but there's some small probability that on some occasion it will have a significant impact. And that can make it worthwhile.
I found this reasonable at the time, but I'm now inclined to think that it's a poor generalization where the expected impact still remains negligible in most scenarios. The main influence for my shift is when I think about how decisions are made within organizations, and how power-seeking approaches are vastly superior to voting in most areas of life where the system exceeds a threshold of complexity.
Anyone care to propose updates on this topic?
This position is commonly defended for consequentialist arguments for vegetarianism and veganism; see, e.g., Section 2 here, Section 2 here, and especially Day 2 here. The argument usually goes something like: if you stop buying one person's worth of eggs, then in expectation, the industry will not produce something like one pound of eggs that they would've produced otherwise. Even if you are not the tipping point to cause them to cause production, due to uncertainty you still have positive expected impact. (I'm being a bit vague here, but I recommend reading at least one of the above readings -- especially the third one -- because they make the argument better than I can.)
In the case of animal product consumption, I'm confused what you mean by "the expected impact still remains negligible in most scenarios" -- are you referring to different situations? I agree in principle that if the expected impact is tiny, then we don't have much reason on consequentialist grounds to avoid the behavior, but do you have a particular situation in mind? Can you give concrete examples of where your shift in views applies/where you think the reasoning doesn't apply well?
One of those sources ("Compassion, by the Pound") estimates that reducing consumption by one egg results in an eventual fall in production by 0.91 eggs, i.e., less than a 1:1 effect.
I'm not arguing against the idea that reducing consumption leads to a long-term reduction in production. I'm doubtful that we can meaningfully generalise this kind of reasoning across different specifics as well as distinct contexts without investigating it practically.
For example, there probably exist many types of food products where reducing your consumption only has like a 0.1:1 effect. (It's also reasonable to consider that there are some cases where reducing consumption could even correspond with increased production.) There are many assumptions in place that might not hold true. Although I'm not interested in an actual discussion about veganism, one example of a strong assumption that might not be true is that the consumption of egg is replaced by other food sources that are less bad to rely on.
I'm thinking that the overall "small chance of large impact by one person" argument probably doesn't map well to scenarios where voting is involved, one-off or irregular events, sales of digital products, markets where the supply chain changes over time because there's many ways to use those products, or where excess production can still be useful. When I say "doesn't map well", I mean that the effect of one person taking action could be anywhere between 0:1 to 1:1 compared to what happens when the sufficient number of people simultaneously make the change in decision-making required for a significant shift. If we talk about one million people needing to vote differently so that a decision is reversed, the expected impact of my one vote is always going to be less than 100% of one millionth, because it's not guaranteed that one million people will sway their vote. If there's only a 10% chance of the one million swayed votes, I'd think my expected impact to come out at far less than even 0.01:1 from a statistical model.
Thanks, this makes things much clearer to me.
I agree that this style of reasoning depends heavily on the context studied (in particular, the mechanism at play), and that we can't automatically use numbers from one situation for another. I also agree with what I take to be your main point: In many situations, the impact is less than 1:1 due to feedback loops and so on.
I'm still not sure I understand the specific examples you provide:
* Not exactly a byproduct, since sales of leather increases the revenue from raising a cow.
** This is not accounting for less direct impacts on demand, like influencing others around oneself.
This is because the mechanism used to change levels of production is similar in these cases.
I'm unclear on the exact mechanism and suspect that the anecdote of "the manager sees the reduced demand across an extended period and decides to lower their store's import by the exact observed reduction" is a gross oversimplification of what I would have guessed is a complex system where the manager isn't perfectly rational, may have long periods without review due to contractual reasons, the supply chain lasting multiple parties all with non-linear relationships. Maybe some food supply chains significantly differ at the grower's end, or in different countries. My missing knowledge here is why I don't think I have a good reason to assume generality.
Other animal products
I think your cow leather example highlights the idea that for me threatens simplistic math assumptions. Some resources are multi-purpose, and can be made into different products through different processes and grades of quality depending on the use case. It's pretty plausible that eggs are either used for human consumption or hatching. Some animal products might be more complicated and be used for human consumption or non-human consumption or products in other industries. It seems reasonable for me to imagine a case where decreasing human consumption results in wasted production which "inspires" someone to redirect that production to another product/market which becomes successful and results in increased non-dietary demand. I predict that this isn't uncommon and could dilute some of the marginal impact calculations which are true short-term but might not play out long-term. (I'm not saying that reducing consumption isn't positive expectation, I'm saying that the true variance of the positive could be very high over a long-term period that typically only becomes clear in retrospect.)
Voting
Thanks for that reference from Ord. I stand updated on voting in elections. I have lingering skepticism about a similar scenario that's mathematically distinct: petition-like scenarios. E.g. if 100k people sign this petition, some organization is obliged to respond. Or if enough students push back on a school decision, the school might reconsider. This is kind of like voting except that the default vote is set. People who don't know the petition exists have a default vote. I think the model described by Ord might still apply, I just haven't got my head around this variation yet.
I agree that the simple story of a producer reacting to changing demand directly is oversimplified. I think we differ in that I think that absent specific information, we should assume that any commonly consumed animal product's supply response to changing demand should be similar to the ones from Compassion, by the Pound. In other words, we should have our prior on impact centered around some of the numbers from there, and update from there. I can explain why I think this in more detail if we disagree on this.
Leather example:
Sure, I chose this example to show how one's impact can be diluted, but I also think that decreasing leather consumption is unusually low-impact. I don't think the stories for other animal products are as convincing. To take your examples:
Voting:
I'll need to think about it more, but as with two-candidate votes, I think that petitions can often have better than 1:1 impact.
Animal Charity Evaluators estimates that a plant-based diet spares 105 vertebrates per year. So if you’re vegan for 50 years, that comes out to 5,250 animals saved. If you put even 10% credence in the ACE number, where the counterfactual is zero impact, you’d still be helping over 500 animals in expectation.
Late to the party here but I'd check out Räuker et al. (2023), which provides one taxonomy of AI interpretability work.
Brilliant, thank you. One of the very long lists of interp work on the forum seemed to have everything as mech interp (or possibly I just don't recognize alternative key words). Does the EA AI safety community feel particularly strongly about mech interp or is it just my sample size being too small?
Not a direct answer, but you might find the Interpretability (ML & AI) tag on LW relevant. That's where I found Neel Nanda's longlist of interpretability theories of impact (published Mar-22 so it may be quite outdated), and Charbel-Raphaël's Against Almost Every Theory of Impact of Interpretability responding to it (published Aug-23, so much more current).
Does anyone have quick tips on who to ask for feedback on project ideas?
I feel like the majority of feedback I get falls into:
The truth is, no one can really talk me out of attempting that specific thesis question except by providing a definitive answer to my specific question. And if the only feedback I get is one of the above, then I might as well only ask people who will be encouraging of my ideas and potentially recommend other people that are good to talk to.
It's relatively uncommon for me to get feedback at the level of tweaking my ideas. Is it even worth trying to solicit that given how few people are in a good position to do so?
I find that asking EAs (or anyone, really) for open-ended feedback tends not to yield novel insight by default. EAs tend to have high openness, and as long something passes the bar of "this has plausible theory of change, has no obviously huge downside and is positive EV enough to be worth exploring" , it isn't subject to particularly intense scrutiny. Consider also that you have thought about the problem for days/weeks whereas they've only thought about it for maybe 10-20 minutes
Haven't found a 100% perfect solution, but usually I express my 2-3 most pressing doubts and ask them whether there's a possible solution to those doubts. It scopes the question towards more precise, detailed and actionable feedback.
Alternatively, if the person has attempted a similar project, I would ask them what goals/principles they found most important, and 2-3 things they wish they knew before they'd started.
I think you're mostly asking the wrong people. I give a lot of feedback on my friends work and projects, but we all work in pretty similar areas so I know how to provide meaningful and detailed criticism and what does and doesn't make sense. If you're asking people who aren't in that field there's a good chance any detailed feedback they give you will be not very useful anyway. You also might be asking the wrong people in terms of their willingness to invest time and energy into giving someone detailed feedback because it really does take a not insignificant amount of both.
I can't be super certain these are useful tips since I don't know the exact nature of the project you're talking about, but
You're probably right, I'm asking the wrong people. I don't know if there are many of the right people to ask within EA or outside of EA. The project cause area is mental health / personal development as it relates to well-being measured through life satisfaction. I feel like my potential sources of targeted feedback are highly constrained because:
Have you put together an actual proposal document? Having a well laid out and argued proposal would help professionals who are willing to take the time understand and critique your work, even if it's kind of unconventional.
I also find that more experienced professionals seem to apply route knowledge and apparently seem reluctant to deviate from considering anything outside of what they practice.
To be fair, this is in a lot of cases not down to incuriosity or being too stupid or stubborn (although there are certainly a fair amount of that as well), but because they practice what they do because they believe it's the best choice based on their years of training and experience. This is where a well put together, persuasive proposal could come in handy, it gives people the time and ability to peruse and think about it in a way that verbal conversation or jumbled messages don't.
If you aren't getting support from practitioners it may be a sign that this is better suited for research at this point. Establish the science behind what you're trying to do before seeking implementation. Researcher are, in my experience, generally much more open to out of the box thinking because their risk is lower and potential rewards higher. They're also more used to looking at proposals and giving feedback, so maybe reaching out to academics or doing something like an MA could be a better option for you at this stage if you're sufficiently committed to this project to do it.
I suspect my ideas are somewhat wild compared to normal thinking, and I think it would take other people who have their own wild thoughts to comfortably critique mine.
Just as an aside, I tend to be wary of this kind of thinking and try to self critique more when I find myself going that that line. The reason I say so is that while there are certainly very valuable wild thoughts that are wild enough that "normal thinkers" can't meaningfully engage with them, they're very few and far between and the illegality more frequently tends be the result of the arguments and thinking not being well supported, ideas not being connected well to each other, or significant flaws in the idea or how its presented.
I used to frequently come across a certain acronym in EA, used in a context like "I'm working on ___" or "looking for other people who also use ___". I flagged it mentally as a curiosity to explore later, but ended up forgetting what the acronym was. I'm thinking it might be CFAR, which seems to have meant CFAR workshops? If so, 1) what happened to them, and 2) was it common for people to work through the material themselves, self-paced?
The copyright banner at the bottom of their site extends to 2024 and the Google form for workshop applications hasn't been deactivated.
I got a copy of the CFAR handbook in late 2022 and the intro had an explicit reference to self study - along the lines of 'we have only used this in workshops, we don't know what the results of self study of this material does and it wasn't written for self study'
So I assume self study wasn't common but I may be wrong
I identify as an anti-credentialist in the sense that I believe ideas can (under ideal circumstances) be considered on merit alone, regardless of how unreliable or bad the source of the idea is. Isn't credentialism basically a form of ad hominem attack?
The Fallacy Fork was an amazing read. Thanks for the pointer!