Senior Staff ML Engineer - AI Safety & Evaluation
Location: San Jose, California
About the Role
We re looking for a Senior Staff Engineer to help lead our efforts in designing, building, and evaluating next-generation safety mechanisms for foundation models. You ll guide a team of research engineers focused on scaling safety interventions, building tooling for red teaming and model inspection, and designing robust evaluations that stress-test models in realistic threat scenarios.
What You ll Do
Lead the development of model-level safety defenses to mitigate jailbreaks, prompt injection, and other forms of unsafe or non-compliant outputs
Design and develop evaluation pipelines to detect edge cases, regressions, and emerging vulnerabilities in LLM behavior
Contribute to the design and execution of adversarial testing and red teaming workflows to identify model safety gaps
Support fine-tuning workflows, pre/post-processing logic, and filtering techniques to enforce safety across deployed models
Work with red teamers and researchers to turn emerging threats into testable evaluation cases and measurable risk indicators
Stay current on LLM safety research, jailbreak tactics, and adversarial prompting trends, and help translate those into practical defenses for real-world products
What We re Looking For
5+ years of experience in machine learning or AI systems, with 2+ years in a technical leadership capacity
Experience integrating safety interventions into ML deployment workflows (e.g., inference servers, filtering layers, etc.)
Good understanding of transformer-based models and experience with LLM safety, robustness, or interpretability
Strong background in evaluating model behavior, especially in adversarial or edge-case scenarios
Strong communication skills and ability to drive alignment across diverse teams
Bachelor s, Master s, or PhD in Computer Science, Machine Learning, or a related field