Research Engineer, Frontier Safety Loss of Control, DeepMind
at Google
Location
San Francisco, CA, USA
Compensation
$174k–$252k USD
Type
full time
Posted
3 weeks ago
Tailor your résumé to this role in 30 seconds.
Free account · ATS keyword check · per-job bullet rewrite by Claude.
Job description
The Loss of Control team contributes to a defense in depth against the risk of misaligned AI systems being deployed. We take the possibility of very advanced AI seriously. We don’t think control is a suitable alternative to alignment in the limit of advancing intelligence. But while AI remains effectively monitorable, we think that control is an important part of an overall strategy for building safe AI.
We are looking for a research engineer for the Frontier Safety Loss of Control team within the AGI Safety and Alignment Team based in either San Francisco or London.
In this role, the core responsibility is to help Google prepare for the internal use of potentially misaligned AI systems. That means building defense-in-depth against AI that might persistently pursue goals that users and system developers did not intend.
Artificial intelligence will be one of humanity’s most transformative inventions. At Google DeepMind, we are a pioneering AI lab with exceptional interdisciplinary teams focused on advancing AI development to solve complex global challenges and accelerate high-quality product innovation for billions of users. We use our technologies for widespread public benefit and scientific discovery, ensuring safety and ethics are always our highest priority.
Responsibilities
- Identify potential harms from misaligned agents and develop strategies for detection and prevention.
- Implement technical controls to monitor agent thoughts, behaviour, and respond to mitigate potential harms.
- Integrate various agent behaviour signals from across the organisation to inform response policies.
- Conduct adversarial testing of controls.
- Work with internal product teams to ensure that control systems are adopted over all high-risk AI surfaces.
Minimum qualifications:
- Bachelor's degree in Computer Science, Machine Learning, or a related technical field, or equivalent practical experience.
- 5 years of experience in engineering and agentic assistance, including software development in Python.
- Experience working in a frontier AI research and development environment.
- Experience working in a professional software engineering or research team environment.
- Experience working with technical stakeholders.
- Experience in frontier model risk.
Preferred qualifications:
- Experience of engineering or product design for AI tools or assistants, especially those focused on ML Research and Development (R&D).
- Experience with cybersecurity detection and response.
- Experience with collaborating or leading an applied ML project.
- Experience with Large Language Model (LLM) training and inference.
- Knowledge of AI control, chain-of-thought and other monitoring, faithfulness and monitorability and related research areas.
Applicants in San Francisco: Qualified applications with arrest or conviction records will be considered for employment in accordance with the San Francisco Fair Chance Ordinance for Employers and the California Fair Chance Act.