
Research Scientist (Field)
About Goodfire
Behind our name: Like fire, AI holds the potential for both immense benefit and significant risk. Just as mastering fire transformed human history, we believe the safe and intentional development of AI will shape the future of our species. Our goal is to tame this new fire.
Goodfire is an AI interpretability research company focused on understanding and intentionally designing advanced AI systems. We believe advances in interpretability will unlock the next frontier of safe and powerful foundation models and that deep research breakthroughs are necessary to make this possible.
Everything we do is in service of that mission. We move fast, take ownership, and constantly push to improve. We believe in acting today rather than tomorrow. We care deeply about the success of the organization and put the team above ourselves.
Goodfire is a public benefit corporation headquartered in San Francisco with a team of the world’s top interpretability researchers and engineers from organizations like OpenAI and DeepMind. We’ve raised $57M from investors like Menlo, Lightspeed and Anthropic and work with customers including Arc Institute, Mayo Clinic, and Rakuten.
The role:
We are looking for a Research Scientist to join our team and help develop robust, scalable systems for deploying interpretability techniques on large AI models. You will collaborate closely with our Foundational Team to translate novel interpretability methods into production-ready tools and work on scaling our infrastructure to handle increasingly large models and complex use cases.
Core responsibilities:
- Conduct impactful research in the fields of mechanistic interpretability and model editing.
- Develop novel techniques and algorithms for extracting, analyzing, visualizing, and manipulating the internal representations and decision-making processes of large AI models
- Design and implement scalable, robust systems for applying interpretability and model design techniques at scale
- Perform novel research on a frontier AI models across a variety of modalities, such as language, image, and genomics models
- Stay up-to-date with the latest developments in AI interpretability and model editing research, and contribute to the broader scientific community through open-source projects and community initiatives
- Report on important field learned scientific insights and advancements to our Foundational Team
Who you are:
Goodfire is looking for experienced individuals who embody our values and share our deep commitment to making interpretability accessible. We care deeply about building a team who shares our values:
Put mission and team first
All we do is in service of our mission. We trust each other, deeply care about the success of the organization, and choose to put our team above ourselves.
Improve constantly
We are constantly looking to improve every piece of the business. We proactively critique ourselves and others in a kind and thoughtful way that translates to practical improvements in the organization. We are pragmatic and consistently implement the obvious fixes that work.
Take ownership and initiative
There are no bystanders here. We proactively identify problems and take full responsibility over getting a strong result. We are self-driven, own our mistakes, and feel deep responsibility over what we’re building.
Action today
We have a small amount of time to do something incredibly hard and meaningful. The pace and intensity of the organization is high. If we can take action today or tomorrow, we will choose to do it today.
If you share our values and have at least two years of relevant experience, we encourage you to apply and join us in shaping the future of how we design AI systems.
What we are looking for:
- PhD in Computer Science, Machine Learning, or a related field, or equivalent experience
- Demonstrated research intuition for interpretability and model editing research
- Solid engineering skills, with proficiency in Python and experience with PyTorch or similar deep learning frameworks
- Demonstrated ability to collaborate with cross-functional teams, including product and engineering
- Demonstrated ability to communicate complex research ideas to diverse audiences
- Passion for AI interpretability and a commitment to responsible AI development
Preferred qualifications:
- Postdoctoral experience or industry research experience in interpretability
- Experience working in a fast-paced, early-stage startup environment
- Experience leading research projects and mentoring junior researchers
- Contributions to open-source AI/ML projects or research codebases
This role offers market competitive salary, equity, and competitive benefits. More importantly, you'll have the opportunity to work on groundbreaking technology with a world-class team on the critical path to ensuring a safe and beneficial future for humanity.
The expected salary range for this position is $200,000 - $400,000 USD.
This role reports to our CTO.
Apply for this job
*
indicates a required field