AI Alignment Engineer
Engineeringfull timeRemotemid level$85k-$130k
About This Role
We're seeking an AI Alignment Engineer to join our AI-first recruiting platform and help us build responsible, interpretable AI systems at scale. In this role, you'll design and implement alignment mechanisms, conduct safety evaluations, and ensure our machine learning models behave predictably and ethically across diverse recruiting use cases. You'll work directly with our ML and product teams to embed alignment principles into our core systems, balancing performance with transparency and fairness.
Requirements
- Demonstrate 3+ years of experience in machine learning, AI safety, or related fields
- Proficiency in Python and deep learning frameworks (PyTorch or TensorFlow)
- Strong understanding of alignment techniques including RLHF, constitutional AI, or similar approaches
- Hands-on experience designing and running evaluations for model behavior and safety
- Experience with prompt engineering, fine-tuning, or working with large language models
- Excellent written and verbal communication skills for documenting alignment research and findings
Skills
AI SafetyModel InterpretabilityReinforcement Learning from Human Feedback (RLHF)PyTorchMechanistic InterpretabilityAdversarial TestingPrompt EngineeringEthics & Fairness in ML
