AGI: Additional Perspectives on Capital, Labor, Climate, and Geopolitics

This note explores how Artificial General Intelligence (AGI) intersects with economic systems, labor structures, environmental sustainability, and global power dynamics.

The emergence of AGI could be an event as economically and socially significant as the industrial revolution (likely even more so), since it targets the cognitive realm which underpins virtually all sectors.


Table of Contents


Capital and Wealth Concentration

The pursuit of AGI is occurring within a capitalist framework, raising questions about who will own and benefit from such powerful intelligence. Without intervention, AGI could greatly concentrate wealth and power. Imagine a company or nation controlling an AGI that can out-invent, out-strategize, and out-negotiate any human: it would attain a near-monopoly on innovation and productivity.

Economic models suggest that if AGI can essentially replace human labor and operate at near-zero marginal cost, the traditional relationship between labor and capital shifts dramatically. One study argues that AGI could push the value of human labor to near zero, with capital owners reaping most gains, leading to extreme inequality and a crisis of demand (people can’t earn to buy goods)1. To avoid systemic collapse, ideas like Universal Basic Income (UBI) or public ownership of AI are floated1. In other words, AGI might force a renegotiation of the social contract: if machines produce all wealth, how is that wealth distributed?

We already see how digital technology tends to yield “winner-takes-most” markets (a few big tech companies dominate due to network effects and high fixed / low marginal cost economics). AGI could amplify this. If one company gets an AGI that can drive all sorts of innovations, they could enter any industry and outcompete incumbents. In a sense, an AGI could become the ultimate “productive asset.” Seth Baum’s survey noted 72 active AGI projects2, but it’s likely only a handful have the scale to succeed. If one of those hits gold, the first-mover advantage might be enormous.

This raises concerns of monopolies unlike any seen before: maybe an “AGI-Microsoft” or “AGI-Google” controlling key infrastructure of the economy. Traditional antitrust might not help if the AGI advantage is too decisive. Some, like economist Tyler Cowen, argue that market competition or diffusion will eventually make AGI widespread. But others fear a scenario depicted in sci-fi works where a single megacorporation (or state) essentially has all the advanced AI and thus calls the shots globally (the Tyrell Corp in Blade Runner or Weyland-Yutani in Alien).

Some advocate treating AGI (or its output) as a public utility rather than a private asset, to prevent a dystopia of “AI overlords” economically.

Labor and Employment

AGI is often envisioned as automating not just physical or routine jobs (as AI does now) but also cognitive and creative jobs (potentially any job). This raises the prospect of mass unemployment or a work revolution.

Optimistic scenarios foresee a world where automation leads to a “post-scarcity” economy: humans are freed from drudgery to pursue education, leisure, and creativity, aided by AI, with wealth redistributed via mechanisms like UBI. Some imagine “fully automated luxury communism” where AI and robots provide abundance and society is reoriented to common good.

Pessimistic scenarios worry about technological unemployment: if our economic system isn’t restructured, millions could be jobless and excluded. A neo-feudal outcome where a tiny elite owning AI enjoys extreme wealth while a large underclass is unemployed. The pace matters too: if AGI breakthroughs come rapidly, society might not adapt in time, causing economic shocks3.

Historically, technological unemployment has been mitigated by new job creation (tractor replaces farmers, but new jobs in manufacturing and services appear). The crucial difference with AGI is the fear that all human skills could be eventually matched. If that’s true, then there may simply be fewer jobs needed for humans. Some tasks might always require a human touch (therapy, artisanal craft), but these might be niche.

Human-AI Collaboration

A more optimistic labor scenario: humans might still do a lot of work but augmented by AI, becoming “centaurs” (like how centaur chess teams, human+AI, initially outperformed either alone). Every professional might have AI assistants boosting their productivity drastically. In that case, perhaps we transition into jobs that are more supervisory or creative using AI as a tool. But if the AI gets too good, the human might become the junior partner or even unnecessary.

Work Ethic and Societal Values

Since the industrial era, identity and societal contribution are tied to work. If AGI breaks that link, we might need to find new ways to value people (beyond their economic output). Some propose a shift to an economy of volunteering, creativity, and caregiving being socially rewarded even if not tied to survival via wages. This is a deep cultural shift.

Scandinavia and others are exploring shorter work weeks and decoupling income from full employment, which might become more mainstream if AI shrinks labor demand.

Climate and Environmental Impact

AGI could influence climate change and the environment in two opposite ways.

The Energy Problem

Training advanced AI models today is already energy-intensive: large neural networks require huge computing clusters that consume electricity (often from fossil fuels) and water for cooling data centers4. GPT-3 was estimated to consume 1,287 MWh for training, emitting approximately 550 tons of CO25. If reaching AGI requires hundreds or thousands times more computing, then energy usage could skyrocket.

Critics note that an arms race for AGI could be an environmental nightmare if not powered by renewable energy. One analysis noted AI is “directly responsible for carbon emissions and millions of gallons of water consumption” at data centers4.

AGI as Climate Solution

On the other hand, AGI might become the ultimate tool for solving environmental problems. A superintelligent system could potentially:

  • Design better solar cells
  • Optimize energy grids globally
  • Invent carbon capture methods
  • Model climate with unparalleled accuracy
  • Coordinate large-scale environmental projects

Sam Altman and others have suggested advanced AI will help “fix the climate”6. There is also hope that smarter systems will accelerate the discovery of clean energy or even geoengineering solutions.

There’s an analogy to nuclear tech: it could power cities or destroy them. AGI might likewise either help solve climate change or worsen it, depending on how it’s used and developed. Some suggest making AGI alignment not just about not harming humans, but also valuing the biosphere: aligning AGI with environmental sustainability too.

Global Geopolitics and Security

Nations see leadership in AI as a strategic asset. The advent of AGI could massively shift the balance of power internationally. A country (or alliance) that develops AGI first might gain decisive advantages in economics, military, and technological supremacy.

The AGI Arms Race

This drives a quasi-arms race mentality: the U.S., China, Russia, and EU are all investing heavily in AI. This competition can spur rapid progress but also raises the risk of reduced cooperation and safety shortcuts (rushing to beat rivals could mean less testing or international dialogue). There’s fear of a Thucydides Trap in AI: tensions between an AI-leading superpower and others could escalate conflicts.

We already see moves like the US imposing export controls on advanced chips to China (because those chips are needed for training cutting-edge AI). This is essentially treating AI progress as a matter of national security (akin to restricting nuclear tech). If AGI development continues, such tech restrictions may intensify, potentially leading to an AI “cold war.”

Military AGI

A big concern is military AGI: an AI that could strategize in war, control autonomous weapons, or even launch cyberattacks. An AGI in warfare context might act faster than human decision loops, potentially leading to accidental conflicts if not properly checked.

Autonomous weapons already pose risk of faster conflict escalation (a drone might retaliate in seconds, giving humans little time to intervene). An AGI controlling cyber operations might launch extremely sophisticated attacks or defenses at blinding speed. This challenges strategic stability. Ex-Google CEO Eric Schmidt has warned that AI will disrupt military balance, advocating for dialogues akin to nuclear arms talks.

International Governance

This has led to calls for international agreements: perhaps a global treaty on AGI development akin to nuclear non-proliferation. Some have proposed an “AGI NPT (Non-Proliferation Treaty)” where countries agree to monitor and prevent any single project from running unchecked.

The difficulty is verification and trust. Unlike nukes, AI is soft: you can hide code more easily than a missile silo. This uncertainty can fuel mistrust. In 2023, we saw initial steps like the US and allies discussing common AI principles, and the UK hosting a global AI safety summit.

Global South and Development

If AGI automates manufacturing and services, countries that rely on labor cost advantage could see their development model upended (why outsource work to a low-wage country if an AI can do it cheaper at home?). This could exacerbate global inequality unless there’s technology transfer or new economic models.

If AGI amplifies productivity, ironically it could either flatten differences (since labor cost differences matter less if machines do everything) or increase them (the country/firm with AGI gets all production). If manufacturing becomes fully automated, companies might relocate factories back to their home country (since cheap labor abroad is irrelevant), potentially hurting developing economies.

On a positive note, AGI delivered via cloud could in theory provide expertise anywhere: a small village could have access to the best diagnostics, education, etc., via AI. But will it be accessible or behind paywalls? The digital divide could widen if AGI requires infrastructure only rich countries have.

These concerns suggest that global governance should consider equitable access to AGI’s benefits, perhaps via international organizations ensuring it’s used for UN Sustainable Development Goals.

Capitalism vs. Other Economic Systems

The combination of AGI + capitalism is especially uncertain. Some thinkers argue that advanced AI could either collapse capitalism or turbocharge it.

Collapse scenario: If profit comes at the cost of eliminating consumer incomes (through job loss) and environment, the system unsustainably implodes, necessitating a new system (maybe some form of socialism or resource-based economy). If people can’t earn, they can’t consume; if they can’t consume, profit can’t be realized.

Turbocharge scenario: Companies with AI might find all sorts of new profit avenues. There’s speculation of AI-driven corporations that operate largely autonomously: an AGI “CEO” could optimize a corporation’s every move, potentially outcompeting human-led firms.

AGI and Post-Capitalism

The idea of AGI forcing post-capitalism is intriguing. Marxist theory predicted that at some point, automation would reduce the need for human labor so much that the labor-based economy would crumble, requiring a new mode of distribution. Some modern Marxists see AGI as the final automation that could validate that prediction.

Already we see how productivity gains from automation haven’t translated into less working hours or broadly shared prosperity; often they went to capital owners. Without policy changes, AGI might continue that trend until perhaps the system breaks.

New Economic Models

Some foresee the need for economic redesign via:

  • Wealth redistribution mechanisms (UBI)
  • Data dividends
  • Collectively owned AI cooperatives

For example, if a government created a “national AGI” and distributed its services freely, that’s a very different outcome than AGI under control of a private monopoly. This raises legal and ethical puzzles: Do we treat an AI-run company differently? Do antitrust laws apply if one AGI-enabled company can do the work of ten and underprice all competition?

Human Dignity and Purpose

Beyond material aspects, AGI raises questions of purpose. Work is not only income; it’s identity and meaning for many. If AGI takes over many roles, society will need to adjust how people find purpose.

Historically, industrial automation moved humans to more cognitive jobs. If AGI even handles cognitive and creative tasks, what is left for humans?

Optimistic view: This could herald a renaissance of leisure and art (as utopians in 19th century envisioned machine liberation leading humans to lives of culture, learning, and play). Maybe AGI handles the tedious work and humans focus on human-to-human care, relationships, and pursuits AIs don’t directly fulfill. Even if AIs can simulate companionship, human authenticity might be valued.

Pessimistic view: A crisis of meaning where, in a world where your contributions are not needed, finding fulfillment becomes difficult. Idle populations can suffer psychological issues or social unrest, especially if inequality remains high.

An oft-mentioned need is re-focusing education and culture toward lifelong learning, creativity, and social connection rather than equating worth with productivity, because AGI might decouple those.

Political Systems and Governance

Authoritarian regimes might use advanced AI/AGI to strengthen surveillance and control (AI monitoring of citizens, predictive policing, censorship with AI). A worrying scenario is a totalitarian state powered by AGI that perfectly monitors dissent and manipulates public opinion with deep fakes and targeted propaganda: a 1984-like state with AI as the all-seeing eye.

Conversely, democracies might use AI to enhance citizen services or direct democracy (imagine AI that helps write laws reflecting people’s preferences optimally). So AGI could tilt the balance between open and closed societies depending on who harnesses it better.

AGI as Risk Multiplier

An often overlooked perspective is AGI and existential risk beyond just AI itself: if an AGI is in the hands of a malicious actor, they could use it to develop bioweapons or other catastrophic technologies much faster. So even if AGI itself is aligned, its use as a tool could amplify other risks (like advanced AI designing a super virus that a human terrorist then deploys).

This ties back to governance: we might need global oversight on how AGI is used in sensitive domains.

Policy Implications

AGI is not just a technological milestone; it’s a force multiplier that will interact with every facet of human society and planet. Its arrival (gradual or sudden) could challenge our economic system’s assumptions, shake up labor markets permanently, either degrade or help restore our environment, and reconfigure international power hierarchies.

This is why discussions of AGI increasingly involve not just computer scientists, but economists, sociologists, ethicists, and policymakers. The stakes are as broad as they can be: ensuring that the “general” in AGI means general benefit, not just general capability.

To prepare, some advocate scenario planning and proactive policy:

  • Experiments with UBI or reduced work weeks now
  • Developing global AI governance frameworks early
  • Heavily investing in renewable energy for computing needs
  • Encouraging multi-stakeholder dialogue (private sector, governments, civil society) on AGI’s development

The hope is to avoid being caught off-guard by a technology that could otherwise exacerbate current crises (inequality, climate, conflict) if left solely to market or nationalist forces. Instead, with wise handling, AGI might become a tool that helps solve those crises: essentially, a double-edged sword that we must collectively decide how to wield.


References

  1. Economic analysis of AGI impact on labor value  2

  2. Artificial general intelligence - Wikipedia 

  3. Studies on technological unemployment and pace of change 

  4. Data center energy and water consumption analysis  2

  5. AI carbon footprint studies 

  6. Sam Altman, OpenAI 

Notes mentioning this note


Here are all the notes in this garden, along with their links, visualized as a graph.

AGI: Additional Perspectives on Capital, Labor,...AI and science fictionAI, future imaginaries, and futures studiesArtificial IntelligenceAnti-Dystopia – A Third Way Between Utopia and...ApproachDr. Armin GrunwaldArticle - Future FailArticle – Future of Terranascient Futures Studies...Article - Futures from RuinsArticles on FuturesArtificial General Intelligence (AGI)Artificial Intelligence and future imaginariesArticle – Beyond Capitalist RealismThe Blue Ant TrilogyBooks on FuturesCLA GameCausal layered analysis (cla)Characteristics of Future ImaginariesCollective MemoryThe Collingridge DilemmaCommon SenseComparable and related approachesComparison of national AI strategiesComponents of Future ImaginariesConceptual challenges and inconsistenciesConclusionCondensation of Future ImaginariesCritical Futures StudiesCultural UndercurrentsCyberpunkDark Enlightenment and Neoreactionary MovementDates and eventsDifferentiation from similar termsDigital GardenEarly Approaches to ImaginariesEffect of performativity on future imaginariesFictional Expectations from BeckertFred PolakPaper – From Critique to Cultural RecoveryFuture Imaginaries from CookFuture Imaginaries from Goode and GodheFuture ScenarioFutures StudiesFutures terminologyFutures TriangleFuturesGlobal business networkGoalsImages of the Future from InayatullahImaginaries from Lockton and CandyImaginaries from an anthropological perspectiveImaginariesInteresting Questions in Futures Studies &...IntroductionIntroductions to Futures Thinking & ForesightJargon WatchKey questions in critical futures studiesLegal ForesightLeitbildLiterature on Critical Futures StudiesLongtermismMaster's Thesis – Future ImaginariesMeaning-Making and AIMegatrendsMetamodernism and Futures StudiesMetamodernismMetaphorMethodsMinimum group size for (future) imaginariesMythNele fischer on critical futures studiesNeo-IntegralismNo future is neutralNo such thing as “future-proof”OODA LoopObservations from the examination of futures and...Official futureOutlook – Application of future imaginariesOverton WindowPattern recognitionPolak GamePresent futures from GrunwaldPrimerPull of the FutureQuestions for the Start of a Foresight ProjectRationalistsReal UtopiasResearch QuestionsRoles of Future ImaginariesSTEEPLESarkar gameSituating the ResearcherSocial ImaginariesSociotechnical Imaginaries from JasanoffSohail InayatullahTescrealThe Difference between Present Futures and Future...Book – The Image of the FutureThe JackpotThe L in STEEPLEThe network stateThe Role of Futures in CapitalismThe distinction between imaginaries and future...The paradox of foresightThe vagueness of future imaginariesThinkers on futuresFuture NotesTools for the examination of future imaginariesTowards a definition of Future ImaginariesTraining Practice for FuturistsSorry, but this is not a trend reportTrendsUtopia as Method and Metamodernism: A SynthesisUtopia as MethodUtopiaVerge frameworkVisionWhat good is scientific rigor when nobody gives a...William GibsonMethod: WindtunnelingWorldbuildingCritical FuturesFuture ImaginariesIndex