Week 3 - How to Reduce S-Risks?

chevron-rightHow Can We Reduce S-Risks?arrow-up-righthashtag

Factors influencing future outcomes

  • technological progress, economic dynamics, cooperation problems, political and cultural trends

  • Values of future decision-makers

Capacity building

  • Face great uncertainty about what the future will look like, about how we can best influence it, and about how we can best reduce s-risks

  • How?

    • building a community interested in and knowledgeable about suffering reduction

    • long-term stability of the movement

    • networking with relevant stakeholders

    • developing and refining our ideas

    • expanding our knowledge of how to best reduce suffering

Moral Circle Expansion

  • Concern for suffering irrespective of

    • time

    • species

    • substrate

  • How?

    • Advocacy for animal welfare reforms

    • Advancing clean meat

Suffering-focused ethics

  • Moral reflection and greater clarity on our fundamental values

Politics and Governance

  • Functional political system

    • Avoid harmful political and social dynamics

      • excessive political polarisation

      • Tribalism

    • strengthen democratic governance

  • How?

    • voting reform

    • public service broadcasting

    • deliberative citizen's assemblies

    • compulsory voting

    • Improving norms and standards of political debate that are more charitable and thoughtful

Reducing risks from malevolent actors

  • Malevolence (psychopathy, Machiavellianism, narcissim, sadism)

  • How?

    • Iterated embryo selection or similar technologies

Maintaining the rule of law

  • Researching how laws e.g., regarding threats and extortion can be made more robust

  • Applicability of those laws in radically different contexts

  • How to apply comparable regulation at level of nations, multinational companies or entirely new types of future actors

  • Space Governance

    • Currently free for all i.e., potential for conflict

  • Distribute future technological capacbilties

Research on bargaining and escalating conflicts

  • Agents that threaten to harm other agents, either in an attempt at extortion or as part of an escalating conflict

  • How?

    • research on how to best prevent such negative-sum dynamics and achieve more cooperative outcomes

      • theoretical foundations of game theory and decision theory or on finding the best ways to change the circumstances in which future agents will exist so that escalating conflicts can be avoided

Worst-case AI safety

  • Technical and governance interventions to avoid escalating conflicts involving transformative AI systems, and achieve cooperative AIarrow-up-right

Surrogate goals

  • increase the likelihood that surrogate goalsarrow-up-right will be implemented successfully in future agents

    • would deflect (the disvalue resulting from) threats add to one’s current goals

    • Surrogate goal i.e., goal that one did not initially care about, in the hope that threats will target this surrogate goal rather than what one initially cared about.

chevron-rightBeginner’s guide to reducing s-risksarrow-up-righthashtag

Suffering Risks

  • risks of events that bring about suffering in cosmically significant amounts

  • Cause

    • incidental s-risks

      • unintended consequences of pursuing large-scale goals

      • e.g.,

        • exploitation of future minds for large-scale computations needed for an interstellar civilization, detailed simulations of evolution

        • spreading wildlife throughout the universe without considering the suffering of the organisms involved.

    • agential

      • intentional harm by intelligent beings with influence over many resources

      • e.g., malevolent or retributive agents gaining control over powerful technology, or from AIs that deliberately create suffering.

    • Natural

      • processes that occur without agents’ intervention

      • e.g., future civilizations not prioritizing reducing unnecessary suffering, for reasons similar to the persistence of wild animal suffering on Earth.

  • Solvability

    • researching factors that likely exacerbate these three mechanisms (especially emerging technologies, social institutions, and values)

      • Preventing lock-in states

        • e.g., deployment of powerful artificially intelligent (AI) agents, which could lead to a future shaped by goals that permit causing astronomical suffering

          • Solving the problem of alignment of AI systems with human intent does not appear to be sufficient or necessary to prevent s-risks from AI.

    • applying insights from this research (e.g., recommending principles for the safe design of artificial intelligence)

    • building the capacity of future people to prevent s-risks

      • putting future generations in a better position to make use of information they will have about impending s-risks

    • Targeted Solutions

      • research into AI designs that decrease their tendencies towards destructive conflicts or reduce near-miss risks; some forms of decision theory research; promotion of coordination between and security within AI labs; and research modeling s-risk-relevant properties of future civilizations

    • Broader Solutions

      • advocating moral norms against taking risks of large-scale suffering; promoting more stable political institutions that are conducive to compromise; and building knowledge that could be used by future actors who are in positions to prevent s-risks

chevron-rightChapter 7 – Risk factors for S-Risksarrow-up-righthashtag

Risk Factors

  • What?

    • significantly increase either the probability or the severity of a very bad outcome

  • Why?

    • Might not not need to know all effects that a given action will have on specific s-risks

    • If we can identify reliable risk factors, we will be able to derive robust and effective interventions for reducing a broad range of s-risks.

Types of Risk Factors

  • Capacity of human civilisation to create large amounts of suffering in the first place

    • context of powerful new technologies

    • ability to create sentient artificial entities

      • exploitation of large numbers of sentient beings, due to our likely insufficient level of moral consideration for artificial minds

    • large-scale space colonisation

      • potentially multiply the total population size of both human and nonhuman beings

  • Willingness but lack of capacity

    • ineffective political institutions or cooperation problems

  • Indifference to s-risks or Ignorance

  • Hostility

    • Disregard moral concerns

    • leave little room for prudent reflection on s-risks or for mutually beneficial compromises

    • escalating conflicts and even outright war between competing factions

  • Malevolent Actors

    • Characteristics of Malevolence

      • Psychopathy

        • persistent antisocial behavior, impaired empathy, callousness, and impulsiveness

      • Narcissism

        • inflated sense of one’s importance and abilities, an excessive need for admiration, and an obsession with achieving fame or power

      • Machiavellianism

        • manipulating and deceiving others to further one’s own interests, indifference to common norms, and ruthless pursuit of power or wealth.

      • Sadism

        • tendency to derive pleasure from inflicting suffering and pain on others.

    • As a Risk factor

      • erosion of interpersonal trust and coordination

      • escalating conflicts and war

      • reckless behaviour in high-stakes situations

Complimentariness of Risk Factors

  • polarisation and conflict

    • can increase the likelihood that a malevolent individual rises to power

  • A dictatorship under a malevolent leader would

    • likely impede efforts to prevent s-risks

  • Advanced technology

    • potentially multiply the harm caused by malevolent individuals

Last updated