Bio

Participation
3



You can send me a message anonymously here: https://www.admonymous.co/will

Comments
434

Could you say more about "practically possible"?

Yeah. I said some about that in the ACX thread in an exchange with a Jeffrey Soreff here. Initially I was talking about a "maximally informed" forecaster/trader, but then when Jeffrey pointed out that that term was ill-defined, I realized that I had a lower-bar level of informed in mind that was more practically possible than some notions of "maximally informed."

What steps do you think one could have taken to have reached, say, a 70% credence?

Basically just steps to become more informed and steps to have better judgment. (Saying specifically what knowledge would be sufficient to be able to form a forecast of 70% seems borderline impossible or at least extremely difficult.)

Before the election I was skeptical that people like Nate Silver and his team and The Economist's election modeling team were actually doing as good a job as they could have been[1] forecasting who'd win the election and now post-election I still remain skeptical that their forecasts were close to being the best they could have been.

[1] "doing as good a job as they could have been" meaning I think they would have made substantially better forecasts in expectation (lower Brier scores in expectation) if figuring out who was going to win was really important to them (significantly more important than it actually was), and if they didn't care about the blowback for being "wrong" if they made a confident wrong-side-of-maybe forecast, and if they were given a big budget to use to do research and acquire information (e.g. $10M), and if they were highly skilled forecasters with great judgment (like the best in the world but not superhuman (maybe Nate Silver is close to this--IDK; I read his book The Signal and the Noise, but it seems plausible that there could still be substantial room for him to improve his forecasting skill)).

Note that I also made five Manifold Markets questions to also help evaluate my PA election model (Harris and Trump means and SDs) and the claim that PA is ~35% likely to be decisive.

  1. Will Pennsylvania be decisive in the 2024 Presidential Election?
  2. How many votes will Donald Trump receive in Pennsylvania? (Set)
  3. How many votes will Donald Trump receive in Pennsylvania? (Multiple Choice)
  4. How many votes will Kamala Harris receive in Pennsylvania? (Set)
  5. How many votes will Kamala Harris receive in Pennsylvania? (Multiple Choice)

(Note: I accidentally resolved my Harris questions (#4 & #5) to the range of 3,300,000-3,399,999 rather than 3,400,000-3,499,999. Hopefully the mods will unresolve and correct this for me per my comments on the questions.)

This exercise wasn't too useful as there weren't enough other people participating in the markets to significantly move the prices from my initial beliefs. But I suppose that's evidence that they didn't think I was significantly wrong.

Before the election I made a poll asking "How much would you pay (of your money, in USD) to increase the probability that Kamala Harris wins the 2024 Presidential Election by 0.0001% (i.e. 1/1,000,000 or 1-in-a-million)?"

You can see 12 answers from rationalist/EA people after submitting your answer to the poll or jumping straight to the results.

I think elections tend to have low aleatoric uncertainty, and that our uncertain forecasts are usually almost entirely due to high epistemic uncertainty. (The 2000 Presidential election may be an exception where aleatoric uncertainty is significant. Very close elections can have high aleatoric uncertainty.)

I think Trump was actually very likely to win the 2024 election as of a few days before the election, and we just didn't know that.

Contra Scott Alexander, I think betting markets were priced too low, rather than too high. (See my (unfortunately verbose) comments on Scott's post Congrats To Polymarket, But I Still Think They Were Mispriced.)

I think some people may have reduced their epistemic certainty significantly and had justified beliefs (not overconfident beliefs) that Trump was ~65-90% likely to win.

I totally am willing to believe that the French whale was not one of those people and actually just got lucky.

But I do think that becoming informed enough to rationally obtain a credence of >65% Trump was practically possible.

Thanks for writing up this post, @Eric Neyman . I'm just finding it now, but want to share some of my thoughts while they're still fresh in my mind before next election season.

This means that one extra vote for Harris in Pennsylvania is worth 0.3 μH. Or put otherwise, the probability that she wins the election increases by 1 in 3.4 million

My independent estimate from the week before the election was about 1 in 874,000.

My methodology was to forecast the number of votes that Harris and Trump would each receive in PA, calculate the probability of a tie in PA given my probability distributions for the number of votes they would each get, then multiply the probability of the PA tie by the probability that PA is decisive (conditional on a tie).

I used normal distributions to model the expected number of votes Harris and Trump would each get for simplicity so I could easily model the outcomes in Google Sheets (even though my credence/PDF did not perfectly match a normal distribution). These were my parameters:

Harris MeanHarris SDTrump MeanTrump SD
345000080000348000090000

Simulating 10,000 elections in Google Sheets with these normal distributions found that about 654 elections per 10,000 were within 20,000 votes, which translates to a 1 in ~306,000 chance of PA being tied. I then multiplied this by a ~35%[1] chance that PA would be decisive (conditional on it being tied), to get a 1 in ~874,000 chance of an extra vote for Harris in PA changing who won overall.

99% of the votes in PA are in right now, with the totals currently at: 3,400,854 for Harris and 3,530,234 for Trump.

This means that the vote totals for Harris and Trump are both within 1 standard deviation of my mean expectation. Harris is about half an SD low and Trump was about half an SD high.

From this, it's not clear that my SDs of 80,000 votes for Harris and 90,000 votes for Trump were too narrow, as your (or Nate's) model expected.

So I think my 1 in ~874,000 of an extra vote for Harris determining the president might have been more reasonable than your 1 in ~3.4 million.

[1] Note: I mistakenly privately thought models and prediction markets might be wrong about the chance of PA being decisive, and thought that maybe it was closer to ~50% rather than ~25-35%, but the reason I thought this was bad and I didn't realize until after the election: I made a simple pattern matching mistake by assuming "the election is a toss-up" meant "it will be a close election". I failed to consider other possibilities like "the election will not be close, but we just don't know which side will win by a lot." (In retrospect, this was a very silly mistake for me to make, especially since I had seen that as of some late-October date The Economist said that the two most likely outcomes of the 128 swing-state combinations was 20% that Trump swings all seven and 7% that Harris swings all 7.)

(copying my comment from Jeff's Facebook post of this)

I agree with this and didn't add it (the orange diamond or 10%) anywhere when I first saw the suggestions/asks by GWWC to do so for largely the same reasons as you.

I then added added it to my Manifold Markets *profile* (not name field) after seeing another user had done so. I didn't know who the user was and didn't know that they had any affiliation with effective giving or EA, and appreciated learning that, hence why I decided to do the same. I'm not committed to this at all and may remove it in the future. https://manifold.markets/WilliamKiely

I have not added it to my EA Forum name or profile. Everyone on the EA Forum knows that a lot of people there engage in effective giving, with a substantial percentage giving 10% or more. And unlike the Manifold Markets case where it was a pleasant surprise to learn that a person gives 10+% (where presumably a much lower percentage of people do than the EA Forum), I don't particularly care to know whether a person on he EA Forum does effective giving or not. I also kind of want to avoid actively proselytizing for it on the EA Forum, since for a lot of people trying to save money and focus on finding direct work instead may be a more valuable path than giving a mere 10% of a typical income at the expense of saving faster.

I have not added it anywhere else besides my Manifold Markets profile as far as I recall.

These new systems are not (inherently) agents. So the classical threat scenario of Yudkowsky & Bostrom (the one I focused on in The Precipice) doesn’t directly apply. That’s a big deal.

It does look like people will be able to make powerful agents out of language models. But they don’t have to be in agent form, so it may be possible for first labs to make aligned non-agent AGI to help with safety measures for AI agents or for national or international governance to outlaw advanced AI agents, while still benefiting from advanced non-agent systems.

Unfortunately, the industry is already aiming to develop agent systems. Anthropic's Claude noe has "computer use" capabilities, and I recall that Demis Hassabis has also stated that agentsbstr the next area to make big capability gains. I expect more bad news in this direction in the next five years.

Today, Miles Brundage published the following, referencing this post:

A few years ago, I argued to effective altruists (who are quite interested in how long it may take for certain AI capabilities to exist) that forecasting isn’t necessarily the best use of their time. Many of the policy actions that should be taken are fairly independent of the exact timeline. However, I have since changed my mind and think that most policymakers won’t act unless they perceive the situation as urgent, and insofar as that is actually the case or could be in the future, it needs to be credibly explained why that’s the case. See: COVID and the many warnings about pandemic preparedness before that.

That sounds great too. Perhaps both axes labels should be possible and it should just be specified for each question asked.

I like the idea of operationalizing the Agree/Disagree as  probability that the statement is true. So "Agree" is 100%, neutral is 50%, disagree is 0%. In this case, 20% vs 40% means something concrete.

Load more