Explore jobs tagged llm-evaluation to discover openings focused on LLM evaluation, model testing, benchmarking, prompt engineering, and data annotation; this curated list of jobs surfaces roles across organizations that build and maintain evaluation pipelines, human-in-the-loop labeling workflows, metrics-driven benchmarking (BERTScore, ROUGE, accuracy), adversarial testing, and model quality assurance. Use the filtering UI to narrow results by seniority, remote vs on-site, tech stack, and domain to find targeted opportunities, review typical responsibilities and required skills, and apply or save searches to accelerate hiring outcomes — action-oriented insights and long-tail keywords for LLM evaluation roles to improve match quality and search relevance.
No Llm Evaluation jobs posted this month
Check back soon or explore all available positions