
Pangram Labs is hiring an AI Tooling and Evaluation Engineer to join our AI research team. You’ll work directly with AI research scientists and engineers to build internal tools, platforms, and systems that power the model development lifecycle: from data collection and annotation through evaluation and deployment monitoring.
This is a full-stack software engineering role with a research support focus. You will spend most of your time building web applications, APIs, and data pipelines, but you’ll need enough ML fluency to understand what the research team needs and why.
Roles and Responsibilities:
Build evaluation infrastructure for standardized model assessments and sign-off, including dashboards and reporting tools to track performance over time
Maintain Label Studio deployment for data annotation campaigns
Build browser based interfaces for exploratory data analysis on training datasets
Build research support systems to monitor customer usage of the platform
Improve upon and support synthetic data generation workflows
Take the next step in your career journey