Subramanyam Sahoo
Independent AI Safety Researcher
Cuttack, Odisha, India · sahoo2vec@gmail.com
AI Safety Researcher specializing in alignment science and governance, with 2.5+ years of academic research experience and a proven publication record. I work on building AI systems that remain reliably aligned under adversarial conditions — through mechanistic interpretability, adversarial self-play, and governance frameworks that address institutional failure, not just model failure. NIT Hamirpur gold medalist. MARS 4.0 fellow, Cambridge AI Safety Hub.
20+
Publications
7
Hackathons
$16.5K
Research Funding
Gold
NIT Hamirpur Medal
Fellowships & Positions
Apr 2026
CORDA Democracy Fellowship — Open Democracy Institute
Ongoing research on Integrity Disclosures for Generative AI in Democratic Information Environments
Feb 2025–
AI Policy Fellow (Remote) — UC Berkeley (BASIS Fellowship)
Conducting research on governance aspects for the Berkeley AI Safety Initiative
Dec 2025–Feb 2026
MARS 4.0 — Cambridge AI Safety Hub
Mentorship for Alignment Research Students · Submitted 1 paper to RLC 2026 on RL Agents
Aug–Dec 2025
E-SOAR — EleutherAI
Summer of Open AI Research · Prompt Optimization for Verifiable Hallucination Reduction
Apr 2025–
Independent Contractor — Outlier AI
Designing synthetic datasets for RL-style post-training and evaluation under controlled task distributions
Summer 2025
Harvard Technical AI Safety & Harvard AI Policy Fellowships
Dual fellowships awarded for Summer 2025
Jul–Oct 2025
Mentor (Remote) — Paragon Policy Fellowship
AI Policy and Technical AI Governance (TAIG) research
Research Funding
AIM Intelligence AI Safety Compute Grant
South Korea · Apr–Jun 2026 · PI
USD 10,000
Martian — Research Grant
Nov 2025–Feb 2026 · PI · Mechanistic Interpretability
USD 6,000
Apart Research — Research Grant
Oct 2025 · Pilot experiments and preliminary analyses
USD 500
Recent
- ICLR 2026 — AI for Peace Workshop — Oral Presentation · Dial E for Ethical Enforcement
- Accepted: AIM Intelligence AI Safety Compute Grant, South Korea — PI (USD 10,000)
- Harvard Technical AI Safety & Harvard AI Policy Fellowships — awarded Summer 2025
- Apart Lab Studio Internship — accepted following Martian Mechanistic Interpretability Hackathon project
- CBRN AI Risk Research Sprint — 3rd Prize · Molecules Under Watch
- Featured in Bloomberg — Intelligence Symbiosis Manifesto signatory
- Y Combinator Startup School 2026 — accepted, Bangalore, India (April 18, 2026)