Week 6 - Risks from Artificial Intelligence

The case for worrying about risks from artificial intelligence

chevron-rightThe case for taking AI seriously as a threat to humanityarrow-up-righthashtag

Uncertainty

  • Some think it is near

  • Some of them think advanced AI is so distant that there’s no point in thinking about it now.

  • Others are worried that excessive hype about the power of their field might kill it prematurely.

  • And even among the people who broadly agree that AI poses unique dangers, there are varying takes on what steps make the most sense today.

What is AI?

Dangers

  • Misalignment

    • Problems come from the systems being really good at achieving the goal they learned to pursue

      • it’s just that the goal they learned in their training environment isn’t the outcome we actually wanted.

    • Example

      • We tell a system to run up a high score in a video game.

        • We want it to play the game fairly and learn game skills — but if it instead has the chance to directly hack the scoring system, it will do that

      • We develop a sophisticated AI system with the goal of, say, estimating some number with high confidence.

      • The AI realizes it can achieve more confidence in its calculation if it uses all the world’s computing hardware, and it realizes that releasing a biological superweapon to wipe out humanity would allow it free use of all the hardware.

      • Having exterminated humanity, it then calculates the number with higher confidence.

  • Unable to turn it off

    • AI could email copies of itself somewhere where they’ll be downloaded and read, or hack vulnerable systems elsewhere

  • Bad Actors abusing powerful AI

Variables for Progress

  • Compute

  • Algorithms

    • Lots of algorithms that seemed not to work at all turned out to work quite well once we could run them with more computing power.

    • Deep Learning

      • Same approach produces fake newsarrow-up-right or musicarrow-up-right depending on what training data it is fed.

      • And as far as we can discover, the programs just keep getting better at what they do when they’re allowed more computation time — we haven’t discovered a limit to how good they can get.

    • Recursive self-improvement

      • a better engineering AI — to build another, even better AI

Neglect

  • Not every organization with a major AI department has a safety team at all

  • Some have safety teams focused only on algorithmic fairness and not on the risks from advanced systems.

  • The US government doesn’t have a department for AI.

Solvability

  • AI Safety Community

  • AI Safety Research

  • AI Safety Policy

chevron-right[Our World in Data] AI timelines: What do experts in artificial intelligence expect for the future? (Roser, 2023)arrow-up-righthashtag

Forecasts of 1128 people

  • 812 individual AI experts

  • aggregated estimates of 315 forecasters from the Metaculus platform

  • findings of the detailed study by Ajeya Cotra

Uncertainty

  • There is no consensus, and the uncertainty is high

  • Disagreement

    • when human-level AI will be developed

  • Agreement

    • shorter than a century

    • e.g., 50% of experts gave a date before 2061

Invitation

  • Public discourse and the decision-making at major institutions have not caught up with these prospects

Strategies for reducing risks from unaligned artificial intelligence

chevron-rightThe longtermist AI governance landscape: a basic overviewarrow-up-righthashtag

Definitions

  • AI governance

    • Bringing about local and global norms, policies, laws, processes, politics, and institutions (not just governments) that will affect social outcomes from the development and deployment of AI systems.

  • Longtermist AI governance

    • subset of this AI governance work that is motivated by a concern for the very long-term impacts of AI

Spectrum of 'AI Work'

Foundational

Applied

Field-building work

Other ways to categorize the AI governance landscape

  • Shifting existing discussions in the policy space to make them more sensitive to AI x-risk (e.g. building awareness of the difficulty of assuring cutting-edge AI systems)

  • Proposing novel policy tools (e.g. international AI standards)

  • Getting governments to fund AI safety research

  • Shifting corporate behaviour (e.g. the windfall clause)

Suffering risks

chevron-rightWhy s-risks are the worst existential risks, and how to prevent themarrow-up-righthashtag

Definition of S-Risks

  • risks of events that bring about suffering in cosmically significant amounts

  • ā€œsignificantā€

    • vastly exceeding all suffering that has existed on Earth so far

    • relative to expected future suffering

Remember X-Risk Definition

  • One where an adverse outcome would either annihilate Earth-originating intelligent life or permanently and drastically curtail its potential

S-risks are a most often a sub-category of X-risks

  • S-risks are the worst x-risks

  • AI Risks

Importance

  • Example

    • Artificial sentience

      • Also, consider the possibility that the first artificially sentient beings we create, potentially in very large numbers, may be ā€œvoiceless

    • Factory farming on mutliple planets

  • How they might come to be

    • Bad Actors

    • Accidental

    • Unavoidable

      • e.g., as part of a conflict

Tractability

  • Moral Circle Expansion

  • Internetional Cooperation

  • S-Risk Research

Neglectedness

  • Some, but not all, familiar work on x-risk also reduces s-risk

  • s-risk gets much less attention than extinction risk

Last updated