← All jobs · Decagon

Senior Software Engineer, ML Infrastructure

Decagon ·
79
AI-Agency
B88 U65
📍 San Francisco, US 💰 $200K–$400K Senior 6+ yrs
PythonPyTorchKubernetesGPUdistributed systemsLLM inferencemulti-node training
TL;DR

Senior ML Infrastructure Engineer at Decagon building distributed training platforms and inference architecture for conversational AI agents. Focus on LLM fine-tuning systems, multi-provider routing, and scaling ML infrastructure for production.

Apply at Decagon →
you'll be redirected to the company's career page

Job description

About Decagon

Decagon is the leading conversational AI platform empowering every brand to deliver concierge customer experiences.

Our technology enables industry-defining enterprises like Avis Budget Group, Block’s Cash App and Square, Chime, Oura Health, and Hunter Douglas to deploy AI agents that power personalized, deeply satisfying interactions across voice, chat, email, SMS, and every other channel.

We’re building a future where customer experiences are being redefined from support tickets and hold music to faster resolutions, richer conversations, and deeper relationships. We’re proud to be backed by world-class investors who share that vision, including a16z, Accel, Bain Capital Ventures, Coatue, and Index Ventures, along with many others.

We’re an in-office company, driven by a shared commitment to excellence and velocity. Our values — Just Get It Done, Invent What Customers Want, Winner’s Mindset, and The Polymath Principle — shape how we work and grow as a team.

About the Team

The ML Infrastructure team builds the systems that power every stage of Decagon's model lifecycle. We own the platforms for model training, the infrastructure for model evaluation and experimentation, and the routing layer that manages inference across multiple providers.

We work at the intersection of research and production: translating cutting-edge ML models into reliable, scalable systems that run in customer environments. We collaborate closely with Research, Infrastructure, and Product teams to ensure models train efficiently, serve reliably, and deliver exceptional user experiences.

The team values technical rigor, pragmatic decision-making, and building systems that others love to use.

About the Role

We're hiring a Senior ML Infrastructure Engineer to own the platforms powering Decagon's model training and inference. You'll build distributed training systems, design inference architecture across multiple providers, and create the frameworks that let our Research and Product teams ship faster.

This role is for someone who thrives on technical depth, can lead multi-quarter initiatives, and wants to shape the long-term architecture of our ML stack.


In this role, you will


Your background looks something like this


Compensation

$200K – $400K + Offers Equity

This range reflects the expected compensation for this role. Compensation within the range is determined based on experience, skills, and the scope of responsibilities, with flexibility for candidates who demonstrate exceptional impact.

In addition to base salary, we offer competitive equity. Final compensation may vary based on location within the United States.

Benefits

We proudly offer the following benefits for our full-time employees:

These benefits are described in more detail in Decagon’s policies, may vary by location, and can change at any time according to applicable compensation and benefits plans.

Apply at Decagon →

More open roles at Decagon

Decagon ·
Senior Software Engineer, Agents
📍 San Francisco, US 💰 $200K–$400K 🛠 AI tools welcome at work · Senior
Senior Software Engineer at Decagon building AI agents for enterprise customer support. Focus on designing and deploying mission-critical agents across voice, chat, and email channels, with direct customer collaboration and platform evolution.
PythonTypeScriptmulti-modal models
83
AI-core
Decagon ·
Staff Software Engineer, ML Infrastructure
📍 San Francisco, US 💰 $200K–$400K · Staff
Staff ML Infrastructure Engineer at Decagon building distributed training platforms and inference architecture for conversational AI. Focus on LLM fine-tuning systems, multi-provider routing, and scaling ML infrastructure for production.
PythonPyTorchKubernetesGPUdistributed systemsLLM inference
81
AI-core
Decagon ·
Engineering Manager, Agent Orchestration
📍 San Francisco, US 💰 $280K–$430K 🛠 AI tools welcome at work · Manager
Engineering Manager at Decagon leading the Agent Orchestration team, responsible for the runtime and control plane powering conversational AI agents. Focus on team leadership, technical strategy, and ensuring reliability and correctness across millions of customer interactions.
distributed systemsexecution enginesagent frameworksorchestration logic
81
AI-core
Decagon ·
Staff Research Engineer, Voice
📍 San Francisco, US 💰 $350K–$475K · Staff
Staff Research Engineer at Decagon building models and algorithms for real-time voice agents. Focus on speech understanding, naturalness, turn-taking, and production deployment of conversational AI.
PythonLLMspeech recognitionmodel trainingevaluation
81
AI-core
Decagon ·
Staff Research Engineer
📍 San Francisco, US 💰 $350K–$475K · Staff
Staff Research Engineer at Decagon building conversational AI models for enterprise support agents. Focus on model training, evaluation, orchestration, and production deployment of LLM-based systems.
PythonLLMPyTorchJAXTensorFlow
81
AI-core