About the job
The AI Platform team at Datadog is at the forefront of innovation, developing the infrastructure that fuels next-generation generative AI features across our diverse product range.
As a Senior Software Engineer on the Evaluation and Annotation team, you will play a pivotal role in designing and enhancing the systems that define and evaluate AI quality at scale. Your responsibilities will include creating evaluation pipelines, monitoring model performance, and establishing annotation workflows that assess correctness, safety, bias, and reliability in real-world applications.
Your contributions will directly influence how Datadog develops and sustains reliable AI capabilities. You will collaborate closely with product, machine learning, and infrastructure teams to set quality standards, integrate evaluation systems with our observability platform, and build human-in-the-loop feedback mechanisms that ensure continuous improvement of model behavior.
At Datadog, we cherish our office culture, valuing the relationships we build, the creativity we foster, and the collaboration that arises from working together. Our hybrid workplace model enables employees to cultivate a work-life balance that suits their individual needs.

