Research Scientist, Cybermonitoring
Snapshot
Artificial Intelligence could be one of humanity’s most useful inventions. At Google DeepMind, we’re a team of scientists, engineers, machine learning experts and more, working together to advance the state of the art in artificial intelligence. We use our technologies for widespread public benefit and scientific discovery, and collaborate with others on critical challenges, ensuring safety and ethics are the highest priority.
About Us
As part of the Security & Privacy Research Team at Google DeepMind, you will play a key role in creating innovative defensive and offensive techniques to protect Gemini and other GenAI models!
The Role
GenAI models and agents are increasingly used to handle sensitive data and permissions alongside untrusted data. For them to operate in a secure, trustworthy, and reliable manner, there are many unsolved, impactful research problems including but not limited to:
-
Adversarially-robust reasoning, coding, and tool-use capabilities under prompt injection and jailbreak attacks.
-
Adherence to privacy norms, whether or not under adversarial prompting.
-
Adversarial techniques against generative models through multi-modal inputs.
-
New model architectures that are secure-by-design against prompt injections.
-
Detecting attacks in the wild and finding ways to mitigate them
Key responsibilities
There are many ways you can drive privacy & security research all the way from ideation, experimentation, to transformative landed impact:
-
Identify unsolved, impactful privacy & security research problems, inspired by the needs of protecting frontier capabilities. Research novel solutions through related work studies, offline and online experiments, and building prototypes and demos.
-
Verify the research ideas in the real world by driving and growing collaborations with Gemini teams working in safety, evaluations, and other related areas to land new innovations together.
-
Amplify the impact by generalizing solutions into reusable libraries and frameworks for protecting Gemini and product models across Google, and by sharing knowledge through publications, open source, and education.
About You
In order to set you up for success as a Research Scientist, Security & Privacy Research at Google DeepMind, we look for the following skills and experience:
-
Ph.D. in Computer Science or related quantitative field, or B.S./M.S. in Computer Science or related quantitative field with 5+ years of relevant experience.
In addition, any of the following would be an advantage:
-
Self-directed engineer/research scientist who can drive new research ideas from conception, experimentation, to productionisation in a rapidly shifting landscape.
-
Strong research experience with LLMs and publications in ML security, privacy, safety, or alignment.
-
Experience with JAX, PyTorch, or similar machine learning platforms.
-
A track record on landing research impact within multi-team collaborative environments under senior stakeholders.
The US base salary range for this full-time position is between $166,000 - $244,000 + bonus + equity + benefits. Your recruiter can share more about the specific salary range for your targeted location during the hiring process.
At Google DeepMind, we value diversity of experience, knowledge, backgrounds and perspectives and harness these qualities to create extraordinary impact. We are committed to equal employment opportunity regardless of sex, race, religion or belief, ethnic or national origin, disability, age, citizenship, marital, domestic or civil partnership status, sexual orientation, gender identity, pregnancy, or related condition (including breastfeeding) or any other basis as protected by applicable law. If you have a disability or additional need that requires accommodation, please do not hesitate to let us know.
Create a Job Alert
Interested in building your career at DeepMind? Get future opportunities sent straight to your email.
Apply for this job
*
indicates a required field