About the AI Safety Institute
The AI Safety Institute (AISI), launched at the 2023 Bletchley Park AI Safety Summit, is the world's first state-backed organization dedicated to advancing AI safety for the public interest. Our mission is to assess and mitigate risks from frontier AI systems, including cyber attacks on critical infrastructure, AI-enhanced chemical and biological threats, large-scale societal disruptions, and potential loss of control over increasingly powerful AI. In just one year, we've assembled one of the largest and most respected model evaluation teams, featuring renowned scientists and senior researchers from leading AI labs such as Anthropic, DeepMind, and OpenAI.
At AISI, we're building the premier institution for impacting both technical AI safety and AI governance. We conduct cutting-edge research, develop novel evaluation tools, and provide crucial insights to governments, companies, and international partners. By joining us, you'll collaborate with the brightest minds in the field, directly shape global AI policies, and tackle complex challenges at the forefront of technology and ethics. Whether you're a researcher, engineer, or policy expert, at AISI, you're not just advancing your career – you're positioned to have significant impact in the age of artificial intelligence.
Autonomous Systems
We're focused on extreme risks from autonomous AI systems - those capable of interacting with the real world. To address this, we're advancing the state of the science in risk modelling, incorporating insights from other safety-critical and adversarial domains, while developing our own novel techniques. We're also empirically evaluating these risks - building out one of the world's largest agentic evaluation suites, as well as pushing forward the science of model evaluations, to better understand the risks and predict their materialisation.
Role Summary
As an AISI resident, you'll be mentored by a multi-disciplinary team including scientists, engineers and domain experts on autonomy risks. You will work in a team of other scholars to build evaluations.
Some of the areas we focus on include:
- Autonomous Research and Development (R&D). Investigating AI systems' potential to conduct research, particularly in studying AI capabilities in developing dual-use technologies, unconventional weapons, and accelerating AI algorithm and hardware development.
- Self-replication. Researching the potential for AI systems to autonomously replicate themselves and persist across networks.
- Deceptive alignment. Evaluating AI systems' potential to deceive human auditors and safety researchers. This includes research into AI's ability and propensity to misrepresent its capabilities, conceal its true objectives, and strategically behave in ways that may not align with its actual goals or knowledge.
You’ll receive coaching from your manager and mentorship from your assigned mentor on our team.
This role has a duration of 6 months. For this role, you must meet the Civil Service nationality rules and have a right to work in the UK.
Skills and Experience Specification
You may be a good fit if you have some of the following skills, experience and attitudes. Please note that you don’t need to meet all of these criteria, and if you're unsure, we encourage you to apply.
- Writing good quality code
- Python
- Frontier model knowledge
- AI safety knowledge
- Research science experience
- Strong written and verbal communication skills.
Salary & Benefits
We are hiring individuals at the L3 level for this role which has an annualised salary of £65,000.
Selection Process
In accordance with the Civil Service Commission rules, the following list contains all selection criteria for the interview process.
Required Experience
We select based on skills and experience regarding the following areas:
- Writing production quality code
- Writing code efficiently
- Python
- Frontier model architecture knowledge
- Frontier model training knowledge
- Model evaluations knowledge
- AI safety research knowledge
- Research problem selection
- Research science
- Written communication
- Verbal communication
- Teamwork
- Interpersonal skills
- Tackle challenging problems
- Learn through coaching
Desired Experience
We additionally may factor in experience with any of the areas that our work-streams specialise in:
- Autonomous systems
- Cyber security
- Chemistry or Biology
- Safeguards
- Safety Cases
- Societal Impacts
Additional Information
Internal Fraud Database
The Internal Fraud function of the Fraud, Error, Debt and Grants Function at the Cabinet Office processes details of civil servants who have been dismissed for committing internal fraud, or who would have been dismissed had they not resigned. The Cabinet Office receives the details from participating government organisations of civil servants who have been dismissed, or who would have been dismissed had they not resigned, for internal fraud. In instances such as this, civil servants are then banned for 5 years from further employment in the civil service. The Cabinet Office then processes this data and discloses a limited dataset back to DLUHC as a participating government organisations. DLUHC then carry out the pre employment checks so as to detect instances where known fraudsters are attempting to reapply for roles in the civil service. In this way, the policy is ensured and the repetition of internal fraud is prevented. For more information please see - Internal Fraud Register.
Security
Successful candidates must undergo a criminal record check and get baseline personnel security standard (BPSS) clearance before they can be appointed. Additionally, there is a strong preference for eligibility for counter-terrorist check (CTC) clearance. Some roles may require higher levels of clearance, and we will state this by exception in the job advertisement. See our vetting charter here.
Nationality requirements
We may be able to offer roles to applicant from any nationality or background. As such we encourage you to apply even if you do not meet the standard nationality requirements (opens in a new window).