All creators
Robert Miles AI Safety

Robert Miles AI Safety

AI Safety, Alignment Theory, and Neural Networks with a focus on technical risk analysis and academic research.

Rating
8.4
ReReview score
Award
Worth Prioritizing
Chart
#6
AI & Software Tools
Subscribers
169K
YouTube
Age
14y 1m
Channel age

Nutrition Label

Robert Miles provides high-fidelity breakdowns of AI safety research, translating dense academic papers into accessible, rigorous explainers. Viewers can expect deep dives into alignment theory, mesa-optimization, and instrumental convergence, often illustrated with precise analogies and code. His content prioritizes educational accuracy and technical nuance over hype.

Strengths

  • +Academic Rigor
  • +Concept Simplification
  • +Precise Terminology

Notes

  • !Citations are rigorous; check video descriptions for direct links to the academic papers and datasets discussed.
  • !Content focuses on theoretical alignment and safety risks rather than consumer product reviews or tutorials.

Why this rating

Evidence receipts showing why each dimension is rated the way it is.

Expertise Signal9/10
Once you have AI systems intelligently pursuing their own goals, you have the first ever technology which isn't just about enabling people to get what they want, but about the technology itself getting what it wants.
[04:40]

Demonstrates precise command of alignment theory concepts (instrumental convergence, agency) and synthesizes them accurately without jargon overload.

Transparency9/10
I didn't take any money from 80,000 Hours for this video, to make sure I would be free to say whatever I wanted.
[14:55]

The creator explicitly discloses the lack of financial incentive to establish independence from the organization being promoted.

Rigor & Evidence9/10
So if the attack policy looks for good opportunities to insert a backdoor and actually goes for it on 0.1% of the problems, there's a probability of 0.85 that it will do it at least once without being spotted by human oversight.
[16:58]

Breaks down the specific statistical logic used in the paper to determine optimal attack rates for the 'Red Team'.

Categories
Automation & AgentsCoding ToolsLLM APIsResearch ToolsSecurity & Privacy
Formats
ExplainersDeep Dives