# AI Safety Research Methods: A Systematic Approach
**Version:** 1.0
**Date:** February 14, 2026
**Purpose:** Rigorous methodology for AI safety research
---
## Why Research Methods Matter
AI safety is a young field with high stakes. Poor research methods lead to:
- Conclusions that don't hold up
- Wasted effort on unpromising approaches
- Missed opportunities for real progress
- Reduced credibility for the field
Good methods lead to:
- Reliable findings
- Efficient progress
- Practical insights
- Field advancement
---
## Core Principles
### Principle 1: Explicit Uncertainty
Always specify confidence levels and acknowledge limitations
### Principle 2: Multiple Perspectives
Consider alternative explanations and steel-man opposing views
### Principle 3: Empirical Grounding
Connect theoretical claims to empirical evidence where possible
### Principle 4: Practical Value
Ensure research has actionable implications
### Principle 5: Reproducibility
Document methods sufficiently for others to replicate
---
## Research Types
### Type 1: Conceptual Analysis
**Purpose:** Clarify concepts, develop frameworks, identify principles
**When to Use:**
- Concept is unclear or contested
- Framework needed for further work
- Theoretical foundation required
**Method:**
```
1. Define scope
- What concept/area?
- What's included/excluded?
- What's the goal?
2. Survey existing work
- Literature review
- Identify key positions
- Map conceptual landscape
3. Analyze
- Identify key distinctions
- Develop framework
- Test against examples
4. Synthesize
- Integrate insights
- Develop framework
- Identify gaps
5. Validate
- Test framework
- Get peer review
- Iterate
```
**Quality Criteria:**
- Clarity: Are concepts well-defined?
- Completeness: Does it cover the domain?
- Usefulness: Can others apply it?
- Rigor: Is reasoning sound?
### Type 2: Literature Review
**Purpose:** Synthesize existing research, identify patterns and gaps
**When to Use:**
- Entering new area
- Mapping field
- Identifying research gaps
**Method:**
```
1. Define scope
- Research questions
- Inclusion/exclusion criteria
- Time frame
2. Search
- Identify databases
- Define search terms
- Apply filters
3. Screen
- Title/abstract review
- Full text review
- Apply criteria
4. Extract
- Key information
- Methodologies
- Findings
5. Analyze
- Identify patterns
- Compare approaches
- Assess quality
6. Synthesize
- Integrate findings
- Identify gaps
- Suggest directions
```
**Quality Criteria:**
- Comprehensiveness: Did you find relevant work?
- Systematic: Was process documented?
- Critical: Did you assess quality?
- Synthesis: Did you integrate findings?
### Type 3: Scenario Analysis
**Purpose:** Explore possible futures, identify risks and opportunities
**When to Use:**
- High uncertainty
- Multiple possible outcomes
- Strategic planning
**Method:**
```
1. Define scope
- What scenarios?
- Time horizon?
- What variables?
2. Identify drivers
- Key factors
- Uncertainties
- Trends
3. Develop scenarios
- Base case
- Best case
- Worst case
- Alternative cases
4. Analyze
- Implications
- Likelihood
- Impact
- Intervention points
5. Validate
- Expert review
- Historical analogies
- Internal consistency
6. Document
- Scenarios
- Reasoning
- Implications
```
**Quality Criteria:**
- Plausibility: Are scenarios realistic?
- Distinctiveness: Do scenarios differ meaningfully?
- Usefulness: Do they inform decisions?
- Rigor: Is reasoning documented?
### Type 4: Framework Development
**Purpose:** Create systematic approaches for analysis or action
**When to Use:**
- Need structured approach
- Complex problem
- Multiple factors
**Method:**
```
1. Identify need
- What problem?
- Why framework needed?
- Who will use it?
2. Survey approaches
- Existing frameworks
- Related fields
- Best practices
3. Design framework
- Core components
- Structure
- Process
4. Test framework
- Apply to cases
- Identify weaknesses
- Iterate
5. Document
- Clear explanation
- Examples
- Guidance
6. Validate
- Peer review
- User testing
- Refinement
```
**Quality Criteria:**
- Completeness: Does it cover the domain?
- Clarity: Can others understand/use it?
- Usefulness: Does it solve the problem?
- Rigor: Is it systematically developed?
### Type 5: Comparative Analysis
**Purpose:** Compare approaches, identify strengths and weaknesses
**When to Use:**
- Multiple approaches exist
- Need to choose between options
- Understanding trade-offs
**Method:**
```
1. Define comparison
- What to compare?
- What criteria?
- What's the goal?
2. Characterize approaches
- Each approach described
- Key features
- Underlying assumptions
3. Evaluate
- Against criteria
- Strengths
- Weaknesses
4. Compare
- Side-by-side
- Trade-offs
- Context dependence
5. Recommend
- When to use each
- Best practices
- Hybrid approaches
```
**Quality Criteria:**
- Fairness: Is each approach represented fairly?
- Comprehensiveness: Are criteria complete?
- Objectivity: Is evaluation unbiased?
- Usefulness: Does it help decisions?
---
## Analysis Methods
### Method 1: INT Prioritization
**Purpose:** Prioritize problems or opportunities
**Process:**
```
1. Score Importance (0-10)
- Scale: How many affected?
- Severity: How bad/good?
- Irreversibility: Can it be undone?
2. Score Neglectedness (0-10)
- Current attention (inverse)
- Funding (inverse)
- Awareness (inverse)
3. Score Tractability (0-10)
- Technical feasibility
- Clarity of path
- Timeline
4. Calculate Priority
Priority = I × N × T
```
### Method 2: COMPLEX Problem Analysis
**Purpose:** Understand and address complex problems
**Components:**
```
C - Context (historical, systems, stakeholders)
O - Objectives (primary, stakeholder, hierarchy)
M - Mechanisms (causal chains, drivers, variables)
P - Patterns (empirical, comparative, anomalies)
L - Leverage Points (types, assessment, intervention)
E - Evidence (types, quality, uncertainty)
X - eXecute (strategy, implementation, learning)
```
### Method 3: Scenario Planning
**Purpose:** Explore possible futures
**Process:**
```
1. Identify driving forces
2. Identify critical uncertainties
3. Develop scenario matrix
4. Flesh out scenarios
5. Analyze implications
6. Identify indicators
```
### Method 4: Root Cause Analysis
**Purpose:** Identify underlying causes
**Process:**
```
1. Define problem
2. Ask "why" 5 times
3. Map causal chain
4. Identify root causes
5. Design interventions
```
### Method 5: Stakeholder Analysis
**Purpose:** Understand who's affected and their interests
**Process:**
```
1. Identify stakeholders
2. Map interests/concerns
3. Assess power/influence
4. Identify conflicts/alignments
5. Design engagement strategy
```
---
## Quality Assurance
### Quality Checklist
**For All Research:**
```
☐ Clear research question
☐ Documented methodology
☐ Multiple perspectives
☐ Confidence levels specified
☐ Limitations acknowledged
☐ Practical implications
☐ Reproducible documentation
```
**For Theoretical Work:**
```
☐ Concepts clearly defined
☐ Assumptions explicit
☐ Reasoning sound
☐ Counterarguments addressed
☐ Examples provided
☐ Scope defined
```
**For Empirical Work:**
```
☐ Data sources documented
☐ Methods reproducible
☐ Limitations acknowledged
☐ Alternative explanations
☐ Statistical rigor
☐ Peer review
```
### Peer Review Process
**Submission:**
```
1. Self-review against checklist
2. Identify specific feedback needs
3. Provide context
4. Submit to reviewer
```
**Review:**
```
1. Overall assessment
2. Specific feedback on:
- Methodology
- Reasoning
- Conclusions
- Presentation
3. Constructive suggestions
4. Quality rating
```
**Revision:**
```
1. Address critical issues
2. Consider all feedback
3. Document changes
4. Resubmit if needed
```
---
## Common Pitfalls
### Pitfall 1: Insufficient Literature Review
**Problem:** Reinventing the wheel, missing key insights
**Solution:** Systematic literature review before starting
### Pitfall 2: Overclaiming
**Problem:** Conclusions exceed evidence
**Solution:** Specify confidence levels, acknowledge limitations
### Pitfall 3: Single Perspective
**Problem:** Only considering one viewpoint
**Solution:** Actively seek alternative perspectives
### Pitfall 4: Missing Practical Value
**Problem:** Theoretical work with no application
**Solution:** Always include practical implications
### Pitfall 5: Poor Documentation
**Problem:** Methods unclear, not reproducible
**Solution:** Document everything thoroughly
---
## Research Workflow
### Phase 1: Planning
**Activities:**
- Define research question
- Determine type and scope
- Plan methodology
- Identify resources
**Output:** Research plan
### Phase 2: Execution
**Activities:**
- Literature review
- Data collection
- Analysis
- Documentation
**Output:** Draft findings
### Phase 3: Review
**Activities:**
- Self-review
- Peer review
- Revision
- Quality check
**Output:** Revised draft
### Phase 4: Publication
**Activities:**
- Final formatting
- Publication decision
- Dissemination
- Feedback collection
**Output:** Published work
---
## Tools and Resources
### Literature Management
- Reference managers (Zotero, Mendeley)
- Search databases (arXiv, Google Scholar)
- Note-taking systems
### Analysis Tools
- Frameworks (INT, COMPLEX)
- Visualization tools
- Statistical software
### Collaboration
- Version control (Git)
- Communication tools
- Shared documents
### Quality Assurance
- Checklists
- Peer review systems
- Quality standards
---
*"Rigorous methods are the foundation of reliable knowledge. In AI safety, where stakes are high and uncertainty is common, methodological rigor is essential."*
**Purpose:** Systematic, rigorous research
**Use:** Guide for all AI safety research
**Outcome:** Reliable, valuable findings