Engineer - Agents & Evals
LovableFull timePosted 24 February 2026
Job Description
TL;DR - We’re looking for strong engineers (backend, frontend, or full-stack) who are excited about building agents. You’ll help shape how we build, evaluate, orchestrate, and scale LLM-powered agents in production - and define what it means to create truly lovable AI products.Why Lovable?Lovable lets anyone and everyone build software with plain English. From solopreneurs to Fortune 100 teams, millions of people use Lovable to transform raw ideas into real products - fast. We are at the forefront of a foundational shift in software creation, which means you have an unprecedented opportunity to change the way the digital world works. Over 2 million people in 200+ countries already use Lovable to launch businesses, automate work, and bring their ideas to life. And we’re just getting started.We’re a small, talent-dense team building a generation-defining company from Stockholm. We value extreme ownership, high velocity and low-ego collaboration. We seek out people who care deeply, ship fast, and are eager to make a dent in the world.What we’re looking forCore requirement: Strong engineering fundamentals. You might be backend, frontend, or full-stack - what matters most is your ability to build high-quality production systems.Big plus if you:Have built AI agents yourself (side projects count)Are deeply curious about how AI systems behave and improveHave worked with LLMs or AI systems in productionAre excited about experimenting with new models and techniquesExperience that stands out:Shipped ML or AI features to real users with uptime requirementsBuilt evaluation systems or ML experimentation pipelinesStrong opinions on safety, latency, and helpfulness - but open to testing and learningWhat you’ll doIn one sentence: Build, tune, and scale agents that power lovable products.Add new agent skills and toolsImprove agent reasoning, orchestration, and efficiencyDesign how multiple agents collaborateSelect the right models for different task typesPush the limits of what agents can reliably do in real productsAnalyze agent behavior and performance“Hill-climb” toward better helpfulness, safety, and reliabilityBuild evaluation frameworks and benchmarksCreate experimentation pipelines and feedback loopsEnsure agents perform well across real-world use casesOur tech stackWe're building with tools that both humans and AI love:React, Golang, and RustServerless primitives on Cloudflare, k8s, GCP, Modal, TemporalHow we hireFill in a short form then jump on an initial exploratory call.Join us for a round of technical interviewsTeam intro calls with the teamAbout your applicationPlease submit your application in English - it’s our company language so you’ll be speaking lots of it if you joinWe treat all candidates equally - if you’re interested please apply through our careers portal
Apply Now
Direct link to company career page