Senior Engineering Manager - Voice AI
Listed on 2026-01-02
-
Software Development
AI Engineer, Cloud Engineer - Software
About Gladly
Uniquely powered by Customer AI, Gladly is the only CX platform that puts the customer — not tickets — at the center of every conversation. Trusted by the world's most customer-centric brands, including Crate & Barrel, Ulta Beauty, and Tumi, Gladly delivers radically efficient and radically personal experiences. Gladly was built for B2C leaders who refuse to compromise on operational efficiency or experience.
With Gladly, every conversation feels uniquely tailored and instantly responsive, emotionally intelligent and operationally precise, deeply personal and effortlessly scalable. At the heart of our simply powerful platform is Gladly Customer AI — uniquely designed to maintain a single, continuous conversation across all channels while striking the perfect balance of technology and human connection. Learn more dly.ai.
We’re a remote‑first company that values flexibility, curiosity, and collaboration. If you're excited about using AI to solve real problems for real people, Gladly offers a unique environment where advances in AI connect directly to products driving customer impact.
About The AI Voice TeamOur mission is to deliver natural, responsive, and reliable AI‑powered voice conversations that match or exceed human agent quality — through the right inference infrastructure, latency optimization, and telephony platform.
This team builds the voice intelligence layer of the platform: real‑time response generation, inference orchestration, and voice system integrations. We’re solving the hardest user experience problems in conversational AI — making sub‑second LLM responses sound natural, keeping conversations flowing during inference latency, and ensuring every AI response is accurate and safe.
Why This Role Is SpecialThis is not just another AI job.
As the Engineering Manager for AI Voice, you'll lead the team that makes AI phone conversations feel human
. You'll architect systems that hide LLM latency, leverage inference providers for reliability and speed, and build closed‑loop feedback systems to make AI responses trustworthy.
We’re not chasing hype — we’re building a production‑grade voice AI platform handling real customer service calls
. This is your chance to have a huge technical and product impact, shaping how AI speaks to millions of customers at scale.
- Lead a product team building the voice AI stack: real‑time inference, latency management, continuous quality monitoring and performance tracking.
- Collaborate closely with Product, Design, AI Core and AI Research teams to define what "great AI voice experience" means — from first greeting to complex issue resolution.
- Drive innovation — stay on the cutting edge of low‑latency LLMs, voice‑optimized models, and real‑time AI architectures.
- Develop and implement engineering processes and best practices.
- Take ownership of projects throughout their full cycle including inception, planning, execution, testing and maintenance.
- Mentor engineers and foster a culture of quality obsession, AI advancement, latency minimization, reliability, and thoughtful system design.
- Engineering leadership experience: 5+ years leading teams delivering real‑time, latency‑sensitive systems at scale.
- Previous experience as a top performing software engineer at an AI or SaaS start‑up.
- LLM infrastructure knowledge: Experience with inference providers, model orchestration, and managing AI backends in production.
- Technical depth: Strong foundation in backend systems, streaming architectures, and low‑latency design patterns.
- Telephony platform knowledge: Experience working with Twilio and WebRTC.
- Bias for action: Proven ability to take ambiguous, early‑stage ideas to production.
- Experience ensuring that all of your engineers with variable backgrounds, seniority and understanding of the domain advance in their career development.
- Collaborative leadership: Excellent communication skills — you mentor engineers and bridge between AI research, infrastructure, and product disciplines.
- Hands‑on experience with low‑latency inference providers, or optimizing LLM response times.
- Voice or conversational AI expertise
: latency hiding, utterance…
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).