On Manifold Markets, "If Artificial General Intelligence has an okay outcome, what will be the reason? — A. Humanity successfully coordinates worldwide to prevent the creation of powerful AGIs for long enough to develop human intelligence augmentation, uploading, or some other pathway into transcending humanity's window of fragility." has a probability of 9.9%.
View on Manifold Markets →This question is currently tracked on Manifold Markets only. When the same question appears on additional platforms, we compute a cross-platform consensus probability. Learn about our methodology →
The only prediction market aggregator with academically validated cross-platform consensus.