AI Agent Engineer; Machine Learning
Listed on 2025-12-18
-
Software Development
AI Engineer, Machine Learning/ ML Engineer
Location: Town of Poland
Join to apply for the Staff AI Agent Engineer (Machine Learning) role at Zendesk
.
Get AI‑powered advice on this job and more exclusive features.
Job DescriptionThe Agentic Tribe is revolutionizing the chatbot and voice assistance landscape with Gen3, a cutting‑edge AI Agent system that pushes the boundaries of conversational AI. Gen3 is a goal‑oriented, dynamic, and truly conversational system capable of reasoning, planning, and adapting to user needs in real time. Leveraging a multi‑agent architecture and advanced language models, Gen3 delivers personalized and engaging user experiences, moving beyond scripted interactions to handle complex tasks and "off‑script" inquiries with ease.
AboutThe Role
We’re seeking a highly experienced and influential AI Agent Engineer to join our team. In this role, you’ll drive innovation and technical leadership at the forefront of AI technology, focusing on designing, developing, and deploying intelligent, autonomous agents that leverage Large Language Models (LLMs) to streamline operations. You’ll shape the cognitive architecture for our AI‑powered applications, creating systems that can reason, plan, and execute complex, multi‑step tasks, and guide other engineers.
WhatYou’ll Do (Responsibilities)
- Architect, design, and lead the development of robust, stateful, and scalable AI agents using Python and modern agentic frameworks (e.g., Lang Chain, Llama Index), setting technical direction and best practices for engineering teams.
- Strategize and oversee the integration of AI agent solutions with existing enterprise systems, databases, and third‑party APIs to create seamless, end‑to‑end workflows across the product, identifying and mitigating architectural risks.
- Evaluate and select appropriate foundation models and services from third‑party providers (e.g., OpenAI, Anthropic, Google), analyzing their strengths, weaknesses, and cost‑effectiveness for specific use cases.
- Own and drive the entire lifecycle of AI Agent deployment, from concept to production and beyond for large, ambiguous, or highly complex initiatives—collaborating closely with cross‑functional teams, including product leadership and ML scientists.
- Troubleshoot, debug, and optimize complex AI systems, ensuring exceptional performance, reliability, and scalability in production environments, and mentoring other engineers in advanced problem‑solving techniques.
- Define, establish, and continuously improve platforms and methodologies for evaluating AI agent performance, setting key metrics, driving iterative improvements across the organization, and influencing industry best practices.
- Establish and enforce best practices for documentation of development processes, architectural decisions, code, and research findings to ensure comprehensive knowledge sharing and maintainability across the team and wider engineering organization.
- Mentor and guide more junior and mid‑level developers, fostering a culture of technical excellence and continuous learning, and contributing to the growth and career development of others.
- Expert in LLM‑Oriented System Design:
Architecting and designing complex multi‑step, tool‑using agents (e.g., Lang Chain, Autogen). Deep understanding of prompt engineering, context management, and LLM behavior quirks (e.g., hallucinations, determinism, temperature effects). Ability to implement advanced reasoning patterns like Chain‑of‑Thought and multi‑agent communication. - Mastery of Tool Integration & APIs:
Designing and implementing secure and scalable integrations of agents with external tools, databases, and APIs (e.g., OpenAI, Anthropic) in complex execution environments, often involving novel solutions or significant architectural considerations. - Retrieval‑Augmented Generation (RAG):
Designing, building, and optimizing highly performant and robust RAG pipelines with vector databases, chunking, and sophisticated hybrid search techniques. - Leadership in Evaluation & Observability:
Defining, implementing LLM evaluation frameworks and comprehensive monitoring for latency, accuracy, and tool usage across production systems, influencing the observability strategy. - Safety &…
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).