Week 6 - Risks from Artificial Intelligence

The case for worrying about risks from artificial intelligence

The case for taking AI seriously as a threat to humanity

Uncertainty

  • Some think it is near

  • Some of them think advanced AI is so distant that there’s no point in thinking about it now.

  • Others are worried that excessive hype about the power of their field might kill it prematurely.

  • And even among the people who broadly agree that AI poses unique dangers, there are varying takes on what steps make the most sense today.

What is AI?

Dangers

  • Misalignment

    • Problems come from the systems being really good at achieving the goal they learned to pursue

      • it’s just that the goal they learned in their training environment isn’t the outcome we actually wanted.

    • Example

      • We tell a system to run up a high score in a video game.

        • We want it to play the game fairly and learn game skills — but if it instead has the chance to directly hack the scoring system, it will do that

      • We develop a sophisticated AI system with the goal of, say, estimating some number with high confidence.

      • The AI realizes it can achieve more confidence in its calculation if it uses all the world’s computing hardware, and it realizes that releasing a biological superweapon to wipe out humanity would allow it free use of all the hardware.

      • Having exterminated humanity, it then calculates the number with higher confidence.

  • Unable to turn it off

    • AI could email copies of itself somewhere where they’ll be downloaded and read, or hack vulnerable systems elsewhere

  • Bad Actors abusing powerful AI

Variables for Progress

  • Compute

    • not having enough computing power to realize our ideas fully

    • cost of a unit of computing time keeps falling

  • Algorithms

    • Lots of algorithms that seemed not to work at all turned out to work quite well once we could run them with more computing power.

    • Deep Learning

      • Same approach produces fake news or music depending on what training data it is fed.

      • And as far as we can discover, the programs just keep getting better at what they do when they’re allowed more computation time — we haven’t discovered a limit to how good they can get.

    • Recursive self-improvement

      • a better engineering AI — to build another, even better AI

Neglect

  • Not every organization with a major AI department has a safety team at all

  • Some have safety teams focused only on algorithmic fairness and not on the risks from advanced systems.

  • The US government doesn’t have a department for AI.

Solvability

  • AI Safety Community

  • AI Safety Research

  • AI Safety Policy

[Our World in Data] AI timelines: What do experts in artificial intelligence expect for the future? (Roser, 2023)

Forecasts of 1128 people

  • 812 individual AI experts

  • aggregated estimates of 315 forecasters from the Metaculus platform

  • findings of the detailed study by Ajeya Cotra

Uncertainty

  • There is no consensus, and the uncertainty is high

  • Disagreement

    • when human-level AI will be developed

  • Agreement

    • shorter than a century

    • e.g., 50% of experts gave a date before 2061

Invitation

  • Public discourse and the decision-making at major institutions have not caught up with these prospects

Strategies for reducing risks from unaligned artificial intelligence

The longtermist AI governance landscape: a basic overview

Definitions

  • AI governance

    • Bringing about local and global norms, policies, laws, processes, politics, and institutions (not just governments) that will affect social outcomes from the development and deployment of AI systems.

  • Longtermist AI governance

    • subset of this AI governance work that is motivated by a concern for the very long-term impacts of AI

Spectrum of 'AI Work'

Foundational

  • Strategy research

    • aims to identify high-level goals we could pursue that, if achieved, would clearly increase the odds of eventual good outcomes from advanced AI

    • Targeted Strategy Research

      • e.g. ā€œI want to find out how much compute the human brain uses, because this will help me answer the question of when TAI will be developed (which affects what high-level goals we should pursue)ā€

    • Exploratory Strategy Research

      • e.g. ā€œI want to find out what China’s industrial policy is like, because this will probably help me answer a bunch of important strategic questions, although I don't know precisely which onesā€

    • Who's doing it?

  • Tactic's research

    • aims to identify plans that will help achieve high-level goals (that strategy research has identified as a priority)

    • Example

      • AI & Antitrust

        • Plan: proposing ways to mitigate tensions between competition law and the need for cooperative AI development

        • High-level goal which this plan is pursuing: increasing cooperation between companies developing advanced AI.

    • Who's doing it?

Applied

  • Policy development

    • Takes goals (from strategy) and output-plans (from tactics) and translates them into policy recommendations that are ready to be delivered to policymakers

    • Policy Advocacy

      • advocates for policies to be implemented

      • e.g. figuring out who is the best person to make the policy ask, to whom, and at what time

    • Policy Implementation

      • work of actually implementing policies in practice, by civil servants or corporations

    • Examples

      • Committing to not incorporate AI technology into nuclear command, control and communications (NC3), e.g. as advocated for by CLTR in their Future Proof report.

      • Government monitoring of AI development, e.g. as developed in this whitepaper on AI monitoring.

    • Who's doing it?

Field-building work

  • aims to build a field of people who are doing valuable work on it.

  • Examples

    1. Bringing in new people by creating:

      • policy fellowships, such as the OpenPhil Technology Policy Fellowship;

      • online programs or courses to help junior people get synced up on what is happening in AI governance;

      • high quality, broadly appealing intro material that reaches many undergraduates;

      • more scalable research fellowships to connect, support and credential interested junior people.

    2. Making the field more effective by creating:

      • research agendas;

      • ways for senior researchers to easily hire research assistants

  • Who's doing it?

Other ways to categorize the AI governance landscape

  • Shifting existing discussions in the policy space to make them more sensitive to AI x-risk (e.g. building awareness of the difficulty of assuring cutting-edge AI systems)

  • Proposing novel policy tools (e.g. international AI standards)

  • Getting governments to fund AI safety research

  • Shifting corporate behaviour (e.g. the windfall clause)

Suffering risks

Why s-risks are the worst existential risks, and how to prevent them

Definition of S-Risks

  • risks of events that bring about suffering in cosmically significant amounts

  • ā€œsignificantā€

    • vastly exceeding all suffering that has existed on Earth so far

    • relative to expected future suffering

Remember X-Risk Definition

  • One where an adverse outcome would either annihilate Earth-originating intelligent life or permanently and drastically curtail its potential

S-risks are a most often a sub-category of X-risks

  • S-risks are the worst x-risks

  • AI Risks

Importance

  • Example

    • Artificial sentience

      • Also, consider the possibility that the first artificially sentient beings we create, potentially in very large numbers, may be ā€œvoiceless

    • Factory farming on mutliple planets

  • How they might come to be

    • Bad Actors

    • Accidental

    • Unavoidable

      • e.g., as part of a conflict

Tractability

  • Moral Circle Expansion

  • Internetional Cooperation

  • S-Risk Research

Neglectedness

  • Some, but not all, familiar work on x-risk also reduces s-risk

  • s-risk gets much less attention than extinction risk

Last updated