At a Glance
- Tasks: Join our AGI safety monitoring team to develop tools that enhance AI safety.
- Company: Apollo Research, a leader in AI safety and risk management.
- Benefits: Competitive salary, flexible hours, unlimited vacation, and professional development budget.
- Why this job: Make a real-world impact on AI safety while working with cutting-edge technology.
- Qualifications: 2+ years in empirical research with AI systems and strong Python skills.
- Other info: Collaborative team environment with opportunities for rapid growth and responsibility.
The predicted salary is between 100000 - 180000 £ per year.
Application deadline: We accept submissions until 16 January 2026. We review applications on a rolling basis and encourage early submissions.
Join our new AGI safety monitoring team and help transform complex AI research into practical tools that reduce risks from AI. As an applied researcher, you'll work closely with our CEO, monitoring engineers and Evals team software engineers to build tools that make AI agent safety accessible at scale. We are building tools that monitor AI coding agents for safety and security failures. You will join a small team and will have significant ability to shape the team & tech, and have the ability to earn responsibility quickly.
You will like this opportunity if you're passionate about using empirical research to make AI systems safer in practice. You enjoy the challenge of translating theoretical AI risks into concrete detection mechanisms. You thrive on rapid iteration and learning from data. You want your research to directly impact real-world AI safety.
KEY RESPONSIBILITIES- Research & Development
- Systematically collect and catalog coding agent failure modes from realāworld instances, public examples, research literature, and theoretical predictions.
- Design and conduct experiments to test monitor effectiveness across different failure modes and agent behaviors.
- Build and maintain evaluation frameworks to measure progress on monitoring capabilities.
- Iterate on monitoring approaches based on empirical results, balancing detection accuracy with computational efficiency.
- Stay current with research on AI safety, agent failures, and detection methodologies.
- Stay current with research into coding security and safety vulnerabilities.
- Monitor Design & Optimization
- Develop a comprehensive library of monitoring prompts tailored to specific failure modes (e.g., security vulnerabilities, goal misalignment, deceptive behaviors).
- Experiment with different reasoning strategies and output formats to improve monitor reliability.
- Design and test hierarchical monitoring architectures and ensemble approaches.
- Optimize log preāprocessing pipelines to extract relevant signals while minimizing latency and computational costs.
- Implement and evaluate different scaffolding approaches for monitors, including chaināofāthought reasoning, structured outputs, and multiāstep verification.
- Future projects (likely not in the first 6 months)
- Fine-tune smaller openāsource models to create efficient, specialized monitors for highāvolume production environments.
- Design and build agentic monitoring systems that autonomously investigate logs to identify both known and novel failure modes.
- 2+ years of experience conducting empirical research with large language models or AI systems.
- Strong experience with AI coding agents, for example, having extensively used and compared frontier coding agents.
- Experience with LLMāasāaājudge setups.
- Experience designing and running experiments, analyzing results, and iterating based on empirical findings (e.g., prompting, scaffolding, agent design, fineātuning, or RL).
- Strong Python programming skills.
- Demonstrated ability to work independently on openāended research problems.
- Experience with AI evaluation frameworks, in particular Inspect (though other frameworks are relevant as well).
- Familiarity with AI safety concepts, particularly agentārelated risks.
- Familiarity with computer security, e.g., security testing and secure system design.
- Experience fineātuning language models or working with smaller openāsource models.
- Previous work building developer tools or monitoring systems.
- Publications or contributions to AI safety or ML research.
- Experience with production log systems or production log analysis.
We want to emphasize that people who feel they don't fulfill all of these characteristics but think they would be a good fit for the position nonetheless are strongly encouraged to apply. We believe that excellent candidates can come from a variety of backgrounds and are excited to give you opportunities to shine.
WHAT YOU'LL ACCOMPLISH IN YOUR FIRST YEAR- Build a comprehensive failure mode database: Systematically collect and categorize 100+ distinct AI agent failure modes across safety and security dimensions, creating the foundation for our monitoring library.
- Develop and validate monitoring approaches: Create and empirically test monitoring prompts and strategies for key failure categories, establishing clear metrics for monitor performance and building evaluation frameworks to track progress.
- Optimize the monitoring pipeline: Improve log preprocessing and monitor scaffolding to achieve measurable improvements in detection accuracy, false positive rates, and computational efficiency.
- Advance monitoring capabilities: Begin work on advanced approaches such as fineātuned specialized monitors or agentic investigation systems, moving our monitoring from reactive detection toward proactive risk identification.
- Hierarchical monitoring for coding agent security: Design a multiālayer monitoring system for detecting security vulnerabilities introduced by coding agents. Start by cataloging common security failure modes (e.g., hardcoded credentials, SQL injection vulnerabilities, insecure API calls). Build specialized monitors for each category, then create a hierarchical system where fast, efficient firstāpass monitors flag potentially problematic code for deeper investigation by more sophisticated monitors. Validate the system on synthetic test cases and real agent outputs, iterating to optimize the tradeāoff between detection rates and false positives while maintaining subāsecond latency for most monitoring decisions.
- Salary: 100k ā 180k GBP (~135k ā 245k USD).
- Flexible work hours and schedule.
- Unlimited vacation.
- Unlimited sick leave.
- Lunch, dinner, and snacks are provided for all employees on workdays.
- Paid work trips, including staff retreats, business trips, and relevant conferences.
- A yearly $1,000 (USD) professional development budget.
- Start Date: Target of 2ā3 months after the first interview.
- Time Allocation: Fullātime.
- Location: The office is in London, and the building is next to the London Initiative for Safe AI (LISA) offices. This is an ināperson role. In rare situations, we may consider partially remote arrangements on a caseābyācase basis.
- Work Visas: We can sponsor UK visas.
The monitoring team is a new team. Especially early on, you will work closely with Marius Hobbhahn (CEO), Jeremy Neiman (engineer) and others on the monitoring team. You'll also sometimes work with our SWEs, Rusheb Shah, Andrei Matveiakin, Alex Kedrik, and Glen Rodgers to translate our internal tools into externally usable tools. Furthermore you will interact with our researchers, since we intend to be "our own customer" by using our tools internally for our research work.
ABOUT APOLLO RESEARCHThe rapid rise in AI capabilities offer tremendous opportunities, but also present significant risks. At Apollo Research, we're primarily concerned with risks from Loss of Control, i.e., risks coming from the model itself rather than e.g. humans misusing the AI. We're particularly concerned with deceptive alignment / scheming, a phenomenon where a model appears to be aligned but is, in fact, misaligned and capable of evading human oversight. We work on the detection of scheming (e.g., building evaluations), the science of scheming (e.g., model organisms), and scheming mitigations (e.g., antiāscheming and control). We closely work with multiple frontier AI companies, e.g. to test their models before deployment or collaborate on scheming mitigations. Apollo aims for a culture that emphasizes truthāseeking, being goalāoriented, giving and receiving constructive feedback, and being friendly and helpful.
EQUALITY STATEMENTApollo Research is an Equal Opportunity Employer. We value diversity and are committed to providing equal opportunities to all, regardless of age, disability, gender reassignment, marriage and civil partnership, pregnancy and maternity, race, religion or belief, sex, or sexual orientation.
HOW TO APPLYPlease complete the application form with your CV. The provision of a cover letter is neither required nor encouraged. Please also feel free to share links to relevant work samples.
INTERVIEW PROCESSOur multiāstage process includes a screening interview, a takeāhome test (approx. 3 hours), 3 technical interviews, and a final interview with Marius (CEO). The technical interviews will be closely related to tasks the candidate would do on the job. There are no leetcodeāstyle general coding interviews. If you want to prepare for the interviews, we suggest getting familiar with the evaluations framework Inspect, or by building simple monitors for coding agents and running them on your own Claude Code / Cursor / Codex / etc. traffic.
YOUR PRIVACY AND FAIRNESS IN OUR RECRUITMENT PROCESSWe are committed to protecting your data, ensuring fairness, and adhering to workplace fairness principles in our recruitment process. To enhance hiring efficiency, we use AIāpowered tools to assist with tasks such as resume screening. These tools are designed and deployed in compliance with internationally recognized AI governance frameworks. Your personal data is handled securely and transparently. We adopt a humanācentred approach: all resumes are screened by a human and final hiring decisions are made by our team. If you have questions about how your data is processed or wish to report concerns about fairness, please contact us at info@apolloresearch.ai.
Applied Researcher, Monitoring in London employer: COL Limited
Contact Detail:
COL Limited Recruiting Team
StudySmarter Expert Advice š¤«
We think this is how you could land Applied Researcher, Monitoring in London
āØTip Number 1
Get to know the team! Before your interview, do a bit of research on the people you'll be working with. Understanding their roles and backgrounds can help you connect during the conversation and show that you're genuinely interested in the team dynamic.
āØTip Number 2
Prepare for practical discussions. Since this role involves building tools and monitoring systems, think about how you can demonstrate your hands-on experience. Be ready to discuss specific projects or challenges you've tackled in the past that relate to AI safety and coding agents.
āØTip Number 3
Show your passion for AI safety! During your interviews, share why you care about making AI systems safer. Talk about any relevant research or personal projects you've worked on that align with the company's mission. This will help us see your commitment to the field.
āØTip Number 4
Donāt forget to ask questions! Prepare thoughtful questions about the team's goals, challenges, and future projects. This not only shows your interest but also helps you gauge if the role is the right fit for you. Plus, it makes for a more engaging conversation!
We think you need these skills to ace Applied Researcher, Monitoring in London
Some tips for your application š«”
Get Your CV Spot On: Make sure your CV is tailored to the role of Applied Researcher. Highlight your experience with AI coding agents and empirical research, as these are key for us. Keep it clear and concise ā we want to see your skills shine!
Showcase Relevant Work: If you've got any projects or publications related to AI safety or monitoring systems, donāt hold back! Share links to your work samples in the application form. We love seeing what youāve done and how it relates to our mission.
Be Yourself: Weāre all about diversity and unique backgrounds. If you think youād be a good fit, even if you donāt tick every box, go ahead and apply! Let your personality come through in your application ā we want to know the real you.
Apply Early: Since we review applications on a rolling basis, itās best to get your application in sooner rather than later. Donāt wait until the deadline ā weāre excited to see what you bring to the table!
How to prepare for a job interview at COL Limited
āØKnow Your Stuff
Make sure you brush up on your knowledge of AI safety concepts and coding agents. Familiarise yourself with the latest research and methodologies in the field, especially around failure modes and detection mechanisms. This will not only help you answer questions confidently but also show your passion for the role.
āØShowcase Your Experience
Prepare to discuss your previous work with empirical research and large language models. Be ready to share specific examples of experiments you've designed, results you've analysed, and how you've iterated based on findings. This is your chance to demonstrate your hands-on experience and problem-solving skills.
āØGet Technical
Since this role requires strong Python programming skills, be prepared to talk about your coding experience. You might even want to bring along a small project or example of code you've written that relates to monitoring systems or AI safety. This will give you an edge and show your practical abilities.
āØAsk Insightful Questions
During the interview, donāt hesitate to ask questions about the teamās current projects or future goals. This shows your interest in the role and helps you gauge if the company culture aligns with your values. Plus, it gives you a chance to engage with the interviewers and make a memorable impression.