Hide table of contents

Depending of different attitudes towards questions like take-off speed, people argue that with the development of AGI we will face situations of world GDP doubling days/weeks/a few years (with the number of years shriking with each further doubling). Many peoples's timelines here seem to be quite broad, including quite commonly expectations like "AGI within the next 2-3 decades very likely".

How the global world order politically as well as economically will change over the next decades is a quite extensively discussed topic in public as well as academia, with many goals and forecasts made until years like 2050 or 2070 ("climate neutral 2050", "china's economy in 30 years"). Barely is AGI mentioned in economics classes, political research papers and the like, despite its apparent impact of making any politics redundant and throwing over any economic forecasts. If AGI was even significantly less mighty than we think and there was even just a 20% chance of it occuring in the next 3 decades, that should be the number one single factor debated in every single argument on any economic/political topic with medium-length scope.  Why, do you think, is it the case, that AGI is comparatively so rarely a topic there?

My motivated reasoning would immediately come up with explanations along the lines of 

  1. people in these disciplines are just not so much aware of AI developments
  2. any forecasts/plans made assuming short timelines and fast takeoff speeds are useless anyways, so it makes sense to just assume longer timelines
  3. Maybe I am just not noticing the omnipresence of AGI debate in economic/political long-term discourse

@1 seems unreasonable, because as soon as the first AI-economics people would come up with these arguments, if they were reasonable, they would become mainstream

@2 if that assumption was consciously made, I'd expect to hear this more often as side note

@3 hard to argue against, given it assumes I don't see the discourse. But I regularity engage with media/content from the UN on their SDGs, have taken some Economics/IR/Politics electives, try to be a somewhat informed citicien and have friends studying these things, and I barely see AI suddenly speeding up things in any forecasts or discussions

Why might this be the case?
To me it seems like either mainstream academia, global institutions and public discourse heavily miss something or we tech/ea/ai people are overly biased in the actual relevance of our own field (I'm CS student)?

52

0
0

Reactions

0
0
New Answer
New Comment

8 Answers sorted by

There are two totally valid conclusions to draw from the structure you've drawn up: that CS people or EA people are deluded, or that the world at large, including extremely smart people, is extremely bad at handling weird or new things.

Empirics has come to dominate econ (all but 1? of the recent American economic review articles were empiric-focused). Where's the data for a treatment effect of a treatment that hasn't occurred? Forecasts and expert predictions are not very stylish either, compared to "true causal effects." I don't know the best approach to solve this and am open to ideas

I suspect this is because there isn't a globally credible/legible consensus body generating or validating the forecasts, akin to the IPCC for climate forecasts that are made with even longer time horizons.

@1 seems unreasonable, because as soon as the first AI-economics people would come up with these arguments, if they were reasonable, they would become mainstream

1 seems the most plausible to me. Reasonable arguments might eventually become mainstream, but that doesn't mean they would do so immediately. 

In particular (a) there may not be many AI-economics people, so the signal could get lost in the noise and (b) economics journals may tend to favour research that focuses on established topics or that uses clever methodology, rather than topics that are important/valuable.

I agree strongly! It would be interesting to research how economists have looked upon the creation of the internet. I guess that there is in fact little research on how the internet would change the world pre-1990. 

we tech/ea/ai people are overly biased in the actual relevance of our own field (I'm CS student)?

You can just as easily say that global institutions are biased about the relevance of their own fields, and I think that is a good enough explanation: Traditional elite fields (press, actors, lawyers, autocrats) don't teach AI, and so can't influence the development of AGI. To perform the feats of confidence that gains or defends career capital in those fields, or to win the agreement and flattery of their peers, they have to avoid acknowledging that AGI is important, because if it's important, then none of them are important.

But, I think this dam will start to break, generally. Economists know better than the other specializations, they have the background in decision theory to know what superintelligence will mean, and they see what's happening in industry. Military is also capable of sometimes recognizing and responding to emerging risks. They're going to start to speak up, and then maybe the rest of the elite will have to face it.

A lot of good potential answers are discussed in this earlier post's comments. My favored explanation is because AGI is a minority concern even within CS academia, so we shouldn't expect it to have much impact outside.

Can I tweet this list. I think it's really good.

I think it's 1. Plenty of ideas are reasonable but not mainstream. E.g. the idea to not attack Ukraine in Russia. The experts probably don't get into the technical arguments and dismiss AGI as hype.

For the same reason that e.g. net electricity generation from fusion power is not the "number one single factor debated in every single argument on any economic/political topic with medium-length scope": Until it exists, it is fictional – why should everyone focus so much on fictional technology? It remains a narrow, academic field. The difference is that there is actual progress towards fusion.

Comments1
Sorted by Click to highlight new comments since:

I do think it'd be interesting to have an AGI-pilled economist talk to one of the economists that do GWP forecasting to see if they can find cruxes.

Curated and popular this week
Relevant opportunities