
Research Engineer, GenAI Post-Training for Security & Privacy
- London
- Permanent
- Full-time
- Building post-training data and tools to improve model capabilities in the problem areas, evaluating and auto-red teaming, and contributing successful solutions into Gemini and other models.
- Amplifying the impact by generalizing solutions into reusable libraries and frameworks for protecting agents and models across Google, and by sharing knowledge through publications, open source, and education.
- B.S./M.S. in Computer Science or related quantitative field with 5+ years of relevant experience.
- Demonstrated experience in training or fine-tuning generative models to improve capabilities.
- Experience with JAX, PyTorch, or similar machine learning platforms.
- Demonstrated experience in Python through strong artifacts in building readable, scalable, reusable ML software.
- Demonstrated experience in adapting research outputs into impactful model improvements, in a rapidly shifting landscape and with a strong sense of ownership.
- Research experience and publications in ML security, privacy, safety, or alignment.
- Being motivated to solve security and privacy issues with frontier models.