Research Engineer – Alignment
Anthropic
Machine LearningHybridSenior$250K–$400K
About the Role
Anthropic is hiring Research Engineers to work on alignment, interpretability, and safety of large language models. You will work alongside world-class researchers to understand and improve how Claude reasons, represents knowledge, and behaves under adversarial conditions.
This role requires strong ML engineering skills (Python, PyTorch/JAX), experience training or fine-tuning large models, and ideally a background in AI safety research.
Interested in this role?
Sign in to apply with your profile and CV.