Comprehensive analysis of pause/moratorium proposals finding they would provide very high safety benefits if implemented (buying time for safety research to close the growing capability-safety gap) but face critical enforcement and coordination challenges with zero current adoption by major labs. The FLI 2023 open letter garnered 30,000+ signatures but resulted in no actual slowdown, highlighting severe tractability issues despite theoretical effectiveness.
Issues3
QualityRated 72 but structure suggests 100 (underrated by 28 points)
Links15 links could use <R> components
StaleLast edited 377 days ago - may need review
Pause / Moratorium
Policy
Pause / Moratorium
Comprehensive analysis of pause/moratorium proposals finding they would provide very high safety benefits if implemented (buying time for safety research to close the growing capability-safety gap) but face critical enforcement and coordination challenges with zero current adoption by major labs. The FLI 2023 open letter garnered 30,000+ signatures but resulted in no actual slowdown, highlighting severe tractability issues despite theoretical effectiveness.
Pause and moratorium proposals represent the most direct governance intervention for AI safety: deliberately slowing or halting frontier AI development to allow safety research, governance frameworks, and societal preparation to catch up with rapidly advancing capabilities. These proposals range from targeted pauses on specific capability thresholds to comprehensive moratoria on all advanced AI development, with proponents arguing that the current pace of development may be outstripping humanity's ability to ensure safe deployment.
The most prominent call for a pause came in March 2023, when the Future of Life InstituteOrganizationFuture of Life Institute (FLI)Comprehensive profile of FLI documenting $25M+ in grants distributed (2015: $7M to 37 projects, 2021: $25M program), major public campaigns (Asilomar Principles with 5,700+ signatories, 2023 Pause ...Quality: 46/100 (FLI) published an open letter calling for a six-month pause on training AI systems more powerful than GPT-4. Released just one week after GPT-4's launch, the letter garnered over 30,000 signatures, including prominent AI researchers such as Yoshua Bengio and Stuart RussellPersonStuart RussellStuart Russell is a UC Berkeley professor who founded CHAI in 2016 with $5.6M from Coefficient Giving (then Open Philanthropy) and authored 'Human Compatible' (2019), which proposes cooperative inv...Quality: 30/100, as well as technology leaders like Elon MuskPersonElon MuskComprehensive profile of Elon Musk's role in AI, documenting his early safety warnings (2014-2017), OpenAI founding and contentious departure, xAI launch, and extensive track record of predictions....Quality: 38/100 and Steve Wozniak. The letter cited risks including AI-generated propaganda, extreme automation of jobs, and a society-wide loss of control. However, no major AI laboratory implemented a voluntary pause, and the letter's six-month timeline passed without meaningful slowdown in frontier development. As MIT Technology Review noted six months later, AI companies instead directed "vast investments in infrastructure to train ever-more giant AI systems."
The fundamental logic behind pause proposals is straightforward: if AI development is proceeding faster than our ability to make it safe, slowing development provides time for safety work. As Bengio et al. wrote in Science in May 2024, "downside artificial intelligence risks must be managed effectively and urgently if posited AI benefits are to be realized safely." However, implementation faces severe challenges including competitive dynamics between nations and companies, enforcement difficulties, and concerns that pauses might push development underground or to jurisdictions with fewer safety constraints. These proposals remain controversial even within the AI safety community, with some arguing they are essential for survival and others viewing them as impractical or counterproductive.
Risk Assessment & Impact
Dimension
Assessment
Rationale
Confidence
Safety Uplift
High (if implemented)
Would buy time for safety research
High
Capability Uplift
Negative
Explicitly slows capability development
High
Net World Safety
Unclear
Could help if coordinated; could backfire if unilateral
Medium
Lab Incentive
Negative
Labs strongly opposed; competitive dynamics
High
Research Investment
$1-5M/yr
Advocacy organizations; FLI, PauseAI
Medium
Current Adoption
None
Advocacy only; no major labs paused
High
Core Arguments
Case for Pausing
Loading diagram...
Argument
Description
Strength
Safety-Capability Gap
Safety research not keeping pace with capabilities
Strong if gap is real
Irreversibility
Some AI risks may be impossible to reverse once realized
Strong for existential risks
Precautionary Principle
Burden of proof should be on developers to show safety
Philosophically contested
Coordination Signal
Demonstrates seriousness; creates space for governance
Moderate
Research Time
Enables catch-up on interpretability, alignment
Strong
Case Against Pausing
Argument
Description
Strength
Enforcement
Unenforceable without international agreement
Strong
Displacement
Development moves to less cautious actors
Moderate-Strong
Lost Benefits
Delays positive AI applications
Moderate
Talent Dispersion
Safety researchers may leave paused organizations
Moderate
False Security
Pause without progress creates complacency
Moderate
Definition Problems
Hard to define what to pause
Strong
Pause Proposals Analyzed
FLI Open Letter (2023)
Aspect
Detail
Scope
Training systems more powerful than GPT-4
Duration
Six months (renewable)
Signatories
30,000+ including Yoshua Bengio, Elon Musk, Stuart Russell, Steve Wozniak, Yuval Noah Harari
Notable critiques: AI researcher Andrew Ng argued that "there is no realistic way to implement a moratorium" without government intervention, which would be "anti-competitive" and "awful innovation policy." Reid Hoffman criticized the letter as "virtue signaling" that would hurt the cause by alienating the AI developer community needed to achieve safety goals.
PauseAI Movement
Aspect
Detail
Founded
May 2023 in Utrecht, Netherlands by software entrepreneur Joep Meindertsma
Network of local organizations; US chapter led by Holly Elmore, UK by Joseph Miller (Oxford PhD)
Approach
Grassroots activism, protests at AI labs (OpenAI Feb 2024, Anthropic Nov 2024), policy advocacy
Policy Asks
Global pause enforced through international treaty; democratic control over AI development
Key Actions
International protests in May 2024 timed to Seoul AI Safety Summit; protests held in San Francisco, New York, Berlin, Rome, Ottawa, London
Academic Proposals
Proposal
Scope
Mechanism
Compute Caps
Limit training compute
Hardware governance
Capability Gates
Pause at defined capability thresholds
Eval-based triggers
Conditional Pause
Pause if safety benchmarks not met
RSP-like framework
Research Moratoria
Pause specific capability research
Targeted restrictions
Implementation Challenges
Coordination Problems
Challenge
Description
Severity
Potential Solution
International Competition
US-China dynamics; neither wants to pause first
Critical
Treaty with verification
Corporate Competition
First-mover advantages; defection incentives
High
Regulatory mandate
Verification
How to confirm compliance
High
Compute monitoring
Definition
What counts as "frontier" AI
High
Clear technical thresholds
Enforcement Mechanisms
Mechanism
Feasibility
Effectiveness
Notes
Voluntary Compliance
Low
Very Low
No incentive to comply
National Regulation
Medium
Medium
Jurisdictional limits
International Treaty
Low-Medium
High if achieved
Requires major power agreement
Compute Restrictions
Medium
Medium-High
Physical infrastructure trackable
Social Pressure
Medium
Low
Insufficient against strong incentives
Unintended Consequences
Consequence
Likelihood
Severity
Mitigation
Development Displacement
High
High
International coordination
Underground Development
Medium
Very High
Compute monitoring
Safety Researcher Exodus
Medium
Medium
Continued safety funding
Competitive Disadvantage
High
Variable
Coordinated action
Delayed Benefits
High
Medium
Risk-benefit analysis
Historical Precedents
Domain
Intervention
Outcome
Lessons
Nuclear Weapons
Various moratoria and treaties
Partial success; proliferation continued
Verification essential
Human Cloning
Research moratoria
Generally effective
Narrow scope helps
Gain-of-Function
Research pause (2014-2017)
Temporary; research resumed
Pressure to resume
Recombinant DNA
Asilomar conference (1975)
Self-regulation worked initially
Community buy-in crucial
CFCs
Montreal Protocol
Highly successful
Clear harm identification
Lessons for AI Pauses
Narrow scope is more enforceable than broad moratoria
Verification mechanisms are essential for compliance
International coordination requires identifying mutual interests
Community buy-in from researchers enables voluntary compliance
Clear triggering conditions help define when restrictions apply
Scalability Assessment
Dimension
Assessment
Rationale
International Scalability
Unknown
Depends on coordination
Enforcement Scalability
Partial
Compute monitoring possible
SI Readiness
Yes (if works)
Would prevent reaching SI until prepared
Deception Robustness
N/A
External policy; doesn't address model behavior
Strategic Considerations
When Pauses Might Work
Condition
Importance
Current Status
International Agreement
Critical
Very limited
Clear Triggers
High
Undefined
Verification Methods
High
Underdeveloped
Alternative Pathway
Medium
Safety research ongoing
Industry Buy-In
Medium-High
Very low
Alternative Approaches
Alternative
Relationship to Pause
Tradeoffs
Differential Progress
Accelerate safety, not slow capabilities
Competitive with capabilities
Responsible Scaling Policies
Conditional pauses at thresholds
Voluntary; lab-controlled
Compute Governance
Indirect slowdown through resource control
More enforceable
International Coordination
Framework for coordinated pause
Slower to achieve
Quick Assessment
Dimension
Rating
Notes
Tractability
Low
Severe coordination and enforcement challenges; no major lab has voluntarily paused
Effectiveness
Very High (if implemented)
Would directly address timeline concerns by buying time for safety research
Neglectedness
Medium
Active advocacy (FLI, PauseAI); major gap in implementation and enforcement mechanisms
Current Maturity
Early Advocacy
FLI letter catalyzed debate but no binding commitments achieved
Time Horizon
Immediate-Long Term
Could theoretically be implemented quickly but requires international coordination
Key Proponents
FLI, PauseAI, Yoshua Bengio
Grassroots movements and prominent AI researchers
Key Opponents
Major AI Labs, Andrew Ng
Competitive dynamics and concerns about practicality
Risks Addressed
If implemented effectively, pause/moratorium would address:
Risk
Mechanism
Effectiveness
Racing DynamicsRiskRacing DynamicsRacing dynamics analysis shows competitive pressure has shortened safety evaluation timelines by 40-60% since ChatGPT's launch, with commercial labs reducing safety work from 12 weeks to 4-6 weeks....Quality: 72/100
Eliminates competitive pressure
Very High
Safety-Capability Gap
Time for safety research
Very High
Governance Lag
Time for policy development
High
Societal Preparation
Time for adaptation
High
Misalignment PotentialAi Transition Model FactorMisalignment PotentialThe aggregate risk that AI systems pursue goals misaligned with human values—combining technical alignment challenges, interpretability gaps, and oversight limitations.
Prevents deployment of unaligned systems
Very High (during pause)
Limitations
Enforcement Infeasibility: No mechanism to enforce global compliance
Safety-Capability GapAi Transition Model ParameterSafety-Capability GapThis page contains no actual content - only a React component reference that dynamically loads content from elsewhere in the system. Cannot evaluate substance, methodology, or conclusions without t...
Gap width
Buys time for safety research to close gap
Racing DynamicsRiskRacing DynamicsRacing dynamics analysis shows competitive pressure has shortened safety evaluation timelines by 40-60% since ChatGPT's launch, with commercial labs reducing safety work from 12 weeks to 4-6 weeks....Quality: 72/100
Competitive pressure
Eliminates racing if universally implemented
A successfully implemented pause would fundamentally alter AI development timelines, providing potentially crucial time for safety research and governance development. However, partial or unilateral implementation may worsen outcomes by shifting development to less safety-conscious actors.