Conditional on no existential catastrophe, will there be a superintelligence by 2040?
Conditional on no existential catastrophe, will there be a superintelligence by 2040?
➕
Plus
265
Ṁ85k
2040
70%
chance
Get
Ṁ1,000
and
S3.00


Sort by:
2mo

24% -> 74% in just over a year is pretty wild

1y

M$5000 limit order up, come at me YESers.

1y

@Lovre Put up another 10k for you.

predictedYES 1y

@IsaacKing Opened this page expecting it at 32%, not at 40%! :)

predictedNO 1y

@Lovre I already gave you 5k at 32! How about we meet in the middle at 36?

predictedYES 1y

@IsaacKing Make it so.

1y

Superintelligence in the normal person's meaning, and not in Isaac's meaning, will be achieved with chain of thought, which means there is essentially a 100% chance it will not "design and deploy a complicated website" in under a minute; the inference involved will take a lot longer than a minute.

predictedNO 1y

@DavidBolin the first one might use chain of thought, but there could be additional iterations by 2040 that speed things up.

predictedNO 1y

@ErickBall Not that much though.

1y

"Can design and deploy a complicated website such as a Facebook clone from scratch in under a minute."

This resolves NO, and so does the one about 2050.

1y

How will you know there isn't a superintelligence?

predictedNO 1y

@ScroogeMcDuck I can't, so I'll resolve based on my best guess at the time.

1y

If this resolves YES then I don't expect my mana to have any value. It will be a record of my forecasting ability in a universe where humans aren't needed for forecasting. Or indeed anything.

So maybe I should "rationally" bet NO?

1y

@MartinRandall You have bragging rights for being correct, that is part of the value

predictedNO 1y

@ShadowyZephyr Sure, but those bragging rights are worth less. It's like bragging rights for being a great horse trainer after the invention of the car.

1y

If e.g. a nuclear exchange kills 90% of humanity but the 10% remaining still manage to build a superintelligence, does that NA this market because there was a realized existential risk even if it wasn't quite successful? Or YES?

Wondering if the conditional is a formality or if it would actually affect resolution in some edge case.

predictedNO 1y

@Mira Mostly a formality. If society is still doing well enough after a nuclear exchange that people are still using Manifold, I wouldn't consider that an existential threat coming to pass.

Technically there exists some form of S-risk where our eternity of torture includes continued usage of Manifold, and in that case I would resolve this N/A. Seems unlikely though. Here's a comparison market:

1y

What's your definition of superintelligence?

predictedNO 1y

@BairAiushin It'll be very obvious if something qualifies.

predictedYES 1y

@IsaacKing what might be obvious to me might not be obvious to you and vice versa. Give at least one criterion please.

predictedNO 1y

Can get a perfect score on any test designed for humans where such a score is theoretically achievable. Can solve any mathematical problem that we know to be in principle solvable with the amount of computing power it has available. Can pass as any human online if given that human's history of online communications and a chance to talk to them. Can consistently beat humans in all computer games. Can design and deploy a complicated website such as a Facebook clone from scratch in under a minute. Can answer any scientific question more accurately than any human once given a chance to read all of the internet.

predictedYES 1y

@IsaacKing thanks, that seems about right

predictedNO 1y
5mo

the "solve any mathematical problem" is a wildy different claim because knowing something is solvable doesn't quantify how hard it is to solve, and some things could be very hard. All the other things can be in a "very easy to a superintelligence" class, but solve any mathematical problem is unknown to me. would prob bet yes if you didn't have that condition

5mo

@JonathanC just a guess, but I’m guessing this would resolve yes if some single (or perhaps some combination) of the criteria were met, and that particular condition is sufficient but not necessary.

What is this?

What is Manifold?
Manifold is the world's largest social prediction market.
Get accurate real-time odds on politics, tech, sports, and more.
Win cash prizes for your predictions on our sweepstakes markets! Always free to play. No purchase necessary.
Are our predictions accurate?
Yes! Manifold is very well calibrated, with forecasts on average within 4 percentage points of the true probability. Our probabilities are created by users buying and selling shares of a market.
In the 2022 US midterm elections, we outperformed all other prediction market platforms and were in line with FiveThirtyEight’s performance. Many people who don't like trading still use Manifold to get reliable news.
How do I win cash prizes?
Manifold offers two market types: play money and sweepstakes.
All questions include a play money market which uses mana Ṁ and can't be cashed out.
Selected markets will have a sweepstakes toggle. These require sweepcash S to participate and winners can withdraw sweepcash as a cash prize. You can filter for sweepstakes markets on the browse page.
Redeem your sweepcash won from markets at
S1.00
→ $1.00
, minus a 5% fee.
Learn more.
© Manifold Markets, Inc.Terms + Mana-only TermsPrivacyRules