I'm at a deliberate inflection point — transitioning from platform engineering into AI safety. Three paths are in view. Each has a different theory of change and a different ask of my background.
Transitioning into technical AI safety research with a focus on ML security, robustness, and the security of agentic AI systems. Likely pathway: DPhil at Oxford AI Security or an intensive fellowship programme. My statistics training, 10+ years of engineering, and hands-on agentic AI work give me a concrete foundation — and my DevSecOps background brings a security-first lens that is underrepresented in ML research.
Specialising in AI security, ML infrastructure security, or building safety evaluation infrastructure — at labs like Anthropic, DeepMind, or dedicated safety organisations. This path offers faster time-to-impact and fills a genuine talent gap: there are far fewer safety-focused engineers than researchers in the ecosystem. My agentic AI technical leadership translates directly.
Combining technical depth with policy and governance work — either as a researcher at GovAI-type organisations, through policy fellowships, or in grantmaking roles that require technical evaluation capacity. My linguistics background and international social enterprise experience are comparative advantages for cross-jurisdictional AI governance work.
Current thinking leans toward Path 1 or 2 — staying deeply technical while pivoting to safety. The core open question: is the marginal safety contribution greater from an engineer who ships immediately, or from one who invests 2–4 years to become a researcher? If you have a perspective, I want to hear it.
Structured upskilling in AI engineering and safety — building from foundations to production systems.
Structured to take you from foundations to production. Build real RAG pipelines, evaluation frameworks, and agentic systems.
You build
Interactive Q&A system on MCP documentation
You build
Ranked chunking strategy backed by your own evaluation evidence
You build
Evidence-based retrieval strategy with 4 techniques evaluated head-to-head
You build
Golden dataset + multi-method evaluation framework
You build
Deployed production chatbot serving real requests
You build
Self-correcting CRAG system + adaptive multi-tool agent
The goal is to create legible output — to practice the craft, get feedback, find collaborators, test your fit, and improve understanding. This is the operating framework for everything that follows.
Cheap tests require the least effort, time, or resources to reduce your uncertainty. Start very short (<1 hour each), progress to short (1–10 hours), then long (10–100 hours), then very long. Each rung gives a stronger signal that you're a good fit — without sunk-cost commitment.
Very short (<1h each)
Short (1–10h each)
Long (10–100h each)
Very long (<1000h each)
Read / Listen / Watch
Follow your nose through the resources below. Prioritise things that build intuition before depth.
Do Stuff — Create Legible Output
Summarise what you read. Write opinions. Code and do maths. Add to GitHub. Post to EA Forum or LessWrong to get feedback.
Network — Learn, Don't Sell
Reach out to people doing the work you want to do. Learn about their path, get feedback on your understanding, build relationships before you need them.
Apply — Even for Feedback
Every application is a cheap test. A rejection with feedback is worth the effort. Use jobs.80000hours.org and the EA Opportunities board.
Concrete next steps toward AI safety — structured by timeframe and grounded in 80,000 Hours career advising.
Now
Near-term
6 Months
Specific roles identified through 80,000 Hours advising as strong matches for this career transition.
Anthropic
Policy Institute (UK)
UK AI Safety Institute
DeepMind
Redwood Research
Institute for AI Policy and Strategy (IAPS)
Centre for the Governance of AI (GovAI)