Lead QA AI Engineer

Anaplan
Manchester, United KingdomPosted 16 January 2026

Job Description

<div class="content-intro"><p>At Anaplan, we are a team of innovators focused on optimizing business decision-making through our leading AI-infused scenario planning and analysis platform so our customers can outpace their competition and the market.</p> <p>What unites Anaplanners across teams and geographies is our collective commitment to our customers’ success and to our Winning Culture.</p> <p style="padding-left: 40px;">Our customers rank among the who’s who in the Fortune 50. Coca-Cola, LinkedIn, Adobe, LVMH and Bayer are just a few of the 2,400+ global companies who rely on our best-in-class platform.</p> <p style="padding-left: 40px;">Our Winning Culture is the engine that drives our teams of innovators. We champion diversity of thought and ideas, we behave like leaders regardless of title, we are committed to achieving ambitious goals, and we love celebrating<em> </em>our wins – big and small.</p> <p>Supported by operating principles of being strategy-led, <a href="https://www.anaplan.com/careers/">values</a>-based and disciplined in execution, you’ll be inspired, connected, developed and rewarded here. Everything that makes you unique is welcome; join us and let’s build what’s next - together!</p></div><p><span data-preserver-spaces="true">We're pioneering a new role focused exclusively on quality assurance for GenAI systems. As our <strong>Lead </strong></span><strong><span data-preserver-spaces="true"><strong>A</strong>I QA Engineer</span></strong><span data-preserver-spaces="true">, you'll develop testing strategies, evaluation frameworks, and quality metrics specifically designed for LLM-powered applications. This role requires a unique blend of QA expertise, understanding of GenAI behaviour, and automation skills </span><span data-preserver-spaces="true">to ensure our AI features are reliable, accurate, and trustworthy. </span></p> <p><strong><span data-preserver-spaces="true">Your Impact</span></strong></p> <ul> <li><span data-preserver-spaces="true">Design and implement comprehensive testing strategies for GenAI features, including conversational AI, agentic systems, and LLM-powered workflows </span></li> <li><span data-preserver-spaces="true">Develop automated test suites for prompt testing, including regression tests that detect unintended changes in model behaviour </span></li> <li><span data-preserver-spaces="true">Create evaluation frameworks to measure GenAI quality across multiple dimensions (accuracy, relevance, safety, consistency, latency) </span></li> <li><span data-preserver-spaces="true">Build and maintain test datasets and golden examples that represent diverse user scenarios and edge cases </span></li> <li><span data-preserver-spaces="true">Implement monitoring and alerting systems to detect quality degradation in production GenAI features </span></li> <li><span data-preserver-spaces="true">Perform adversarial testing to identify potential failures, hallucinations, biases, or security vulnerabilities in AI systems </span></li> <li><span data-preserver-spaces="true">Collaborate with engineers to define acceptance criteria and quality gates for AI feature releases </span></li> <li><span data-preserver-spaces="true">Develop tools and frameworks that ... (truncated, view full listing at source)