Ege<\/td> | 40 years<\/td><\/tr><\/tbody><\/table><\/figure>
You can see the strength of their disagreements in the graphs below, where they give very different probability distributions over two questions relating to AGI development (note that these graphs are very rough and are only intended to capture high-level differences, and especially aren't very robust in the left and right tails).<\/p>In what year would AI systems be able to replace 99% of current fully remote jobs?<\/strong><\/td><\/tr><\/tbody><\/table><\/figure> Median indicated by small dotted line.<\/figcaption><\/figure>In what year will the energy consumption of humanity or its descendants be 1000x greater than now?<\/strong><\/td><\/tr><\/tbody><\/table><\/figure> Median indicated by small dotted line. Note that Ege's median is outside of the bounds at 2177<\/figcaption><\/figure>
So I invited them to have a conversation about where their disagreements lie, sitting down for 3 hours to have a written dialogue. You can read the discussion below, which I personally found quite valuable.<\/p> The dialogue is roughly split in two, with the first part focusing on disagreements between Ajeya and Daniel, and the second part focusing on disagreements between Daniel/Ajeya and Ege.<\/p> I'll summarize the discussion here, but you can also jump straight in.<\/p> Summary of the Dialogue<\/h1>Some Background on their Models<\/h2>Ajeya and Daniel are using a compute-centric model for their AI forecasts, illustrated by Ajeya's draft AI Timelines report<\/a>, and Tom Davidson's takeoff model<\/a> where the question of \"when transformative AI\" gets reduced to \"how much compute is necessary to get AGI and when will we have that much compute? (modeling algorithmic advances as reductions in necessary compute)\". <\/p>Whereas Ege thinks such models should have a lot of weight in our forecasts, but that they likely miss important considerations and doesn't have enough evidence to justify the extraordinary predictions it makes.<\/p> Habryka's Overview of Ajeya & Daniel discussion<\/h2>- Ajeya thinks translating AI capabilities into commercial applications has gone slower than expected (\"it seems like 2023 brought the level of cool products I was naively picturing in 2021<\/i>\") and similarly thinks there will be a lot of kinks to figure out before AI systems can substantially accelerate AI development.<\/li>
- Daniel agrees that impactful commercial applications have been slower than expected, but also thinks that the parts that made that slow can be automated substantially, and that a lot of the complexity comes from shipping something that can be useful to general consumers, and that for applications internal to the company, these capabilities can be unlocked faster.<\/li>
- Compute overhangs also play a big role in the differences between Ajeya and Daniel's timelines. There is currently substantial room to scale up AI by just spending more money on readily available compute. However, within a few years, increasing the amount of training compute further will require accelerating the semiconductor supply chain, which probably can't be easily achieved by just spending more money. This creates a \"compute overhang\" that accelerates AI progress substantially in the short run. Daniel thinks it's more likely than not that we will get transformative AI before this compute overhang is exhausted. Ajeya thinks that is plausible, but overall it's more likely to happen after, which broadens her timelines quite a bit.<\/li><\/ul>
These disagreements probably explain some but not most of the differences in the timelines for Daniel and Ajeya.<\/p> Habryka's Overview of Ege & Ajeya/Daniel Discussion<\/h2>- Ege thinks that Daniel's forecast leaves very little room for Hoftstadter's law (\"It always takes longer than you expect, even when you take into account Hofstadter's Law\"), and in-general that there will be a bunch of unexpected things that go wrong on the path to transformative AI<\/li>
- Daniel thinks that Hofstadter's law is inappropriate for trend extrapolation. I.e. it doesn't make sense to look at Moore's law and be like \"ah, and because of planning fallacy the slope of this graph from today is half of what it was previously\"<\/li>
- Both Ege and Ajeya don't expect a large increase in transfer learning ability in the next few years. For Ege this matters a lot because it's one of the top reasons why AI will not speed up the economy and AI development that much. Ajeya thinks we can probably speed up AI R&D anyways by making AI that doesn't have transfer as good as humans, but is just really good at ML engineering and AI R&D because it was directly trained to be.<\/li>
- Ege expects that AI will have a large effect on the economy, but has substantial probability on persistent deficiencies that prevent AI from fully automating AI R&D or very substantially accelerating semiconductor progress.<\/li><\/ul>
Overall, whether AI will get substantially better at transfer learning (e.g. seeing an AI be trained on one genre of video game and then very quickly learn to play another genre of video game) would update all participants substantially towards shorter timelines.<\/p> We ended the dialogue with Ajeya, Daniel and Ege by putting numbers on how much various AGI milestones would cause them to update their timelines (with the concrete milestones proposed by Daniel). Time constraints made it hard to go into as much depth as we would have liked, but me and Daniel are excited about fleshing more concrete scenarios of how AGI could play out and then collecting more data on how people would update in such scenarios.<\/p> The Dialogue<\/h1>
|
|
|