Research Engineer / Scientist, Alignment Science, London
New Today
Research Engineer / Scientist, Alignment Science, London
About Anthropic
Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.
About The Role
You want to build and run elegant and thorough machine learning experiments to help us understand and steer the behavior of powerful AI systems. You care about making AI helpful, honest, and harmless, and are interested in the ways that this could be challenging in the context of human-level capabilities. You could describe yourself as both a scientist and an engineer. As a Research Engineer on Alignment Science, you\'ll contribute to exploratory experimental research on AI safety, with a focus on risks from powerful future systems, often in collaboration with other teams including Interpretability, Fine-Tuning, and the Frontier Red Team.
Responsibilities
Representative projects include:
- Testing the robustness of safety techniques by training language models to subvert our safety techniques and evaluating interventions.
- Running multi-agent reinforcement learning experiments to test techniques like AI Debate.
- Building tooling to efficiently evaluate the effectiveness of novel LLM-generated jailbreaks.
- Writing scripts and prompts to produce evaluation questions testing models’ reasoning in safety-relevant contexts.
- Contributing ideas, figures, and writing to research papers, blog posts, and talks.
- Conducting experiments that inform AI safety efforts at Anthropic, including the design and implementation of our Responsible Scaling Policy.
You May Be a Good Fit If You
- Have significant software, ML, or research engineering experience
- Have experience contributing to empirical AI research projects
- Have familiarity with technical AI safety research
- Prefer fast-moving collaborative projects to extensive solo efforts
- Pick up slack, even if it goes outside your job description
- Care about the impacts of AI
Strong Candidates May Also
- Have experience authoring research papers in ML, NLP, or AI safety
- Have experience with LLMs
- Have experience with reinforcement learning
- Have experience with Kubernetes clusters and complex shared codebases
Candidates Need Not Have
- All required skills listed; formal certifications or education credentials are not mandatory
Annual Salary
The expected salary range for this position is: £250,000 - £270,000 GBP
Logistics
Education requirements: At least a Bachelor's degree in a related field or equivalent experience.
Location-based hybrid policy: All staff should be in one of our offices at least 25% of the time; some roles may require more time in the office.
Visa sponsorship: We sponsor visas where possible; if we make you an offer we will support visa efforts with our immigration lawyer.
We encourage you to apply even if you do not believe you meet every single qualification. We value diverse perspectives and believe that strong candidates come from a range of backgrounds.
How We\'re Different
We believe the highest-impact AI research is big science. At Anthropic we work as a single cohesive team on a few large-scale research efforts and value impact over smaller, narrow tasks. We collaborate, communicate, and pursue high-impact work while maintaining an empirical science mindset similar to physics and biology. Strong communication skills are important.
For more insight into our directions, our recent research covers topics such as GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI & Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.
Join Us
Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible hours, and a collaborative office space.
Seniority level
- Entry level
Employment type
- Full-time
Job function
- Engineering and Information Technology
- Industries: Research Services
London, England, United Kingdom
- Location:
- Greater London, England, United Kingdom
- Salary:
- £100,000 - £125,000
- Job Type:
- FullTime
- Category:
- Engineering
We found some similar jobs based on your search
-
New Today
Research Engineer / Scientist, Alignment Science, London
-
Greater London
- Engineering
Research Engineer / Scientist, Alignment Science, London About Anthropic Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a...
More Details -
-
New Today
Research Engineer / Scientist, Alignment Science, London
-
Greater London, England, United Kingdom
-
£100,000 - £125,000
- Engineering
Research Engineer / Scientist, Alignment Science, London About Anthropic Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a...
More Details -
-
7 Days Old
Research Engineer / Scientist, Alignment Science, London
-
London
- Engineering
About Anthropic Anthropic's mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineer...
More Details -