×
Register Here to Apply for Jobs or Post Jobs. X

Lead Data Engineer

Job in 411001, Pune, Maharashtra, India
Listing for: Impetus
Full Time position
Listed on 2026-02-14
Job specializations:
  • IT/Tech
    AI Engineer, Machine Learning/ ML Engineer
Job Description & How to Apply Below
About the Role:

We are looking for a Software Engineer who combines deep data engineering expertise with hands-on experience in Generative AI and Agentic AI system development on AWS Cloud.

This role is ideal for someone who can design, build, and deploy production-grade GenAI workflows integrating LLMs, vector databases, and orchestration frameworks—with the same rigor as a traditional data system.

Key Responsibilities:

- Design and maintain data pipelines and AI data infrastructure on AWS (Glue, Lambda, S3, Redshift, Step Functions, Athena, etc.).
- Develop and deploy LLM-based applications and Agentic AI workflows using frameworks like Lang Chain, Llama Index, or Auto Gen.
- Build RAG (Retrieval-Augmented Generation) pipelines using AWS services (S3 + Bedrock + Sage Maker + Open Search/Vector DB).
- Implement agentic reasoning, tool calling, and orchestration for multi-agent workflows.
- Containerize and deploy AI services using Docker, ECS, or EKS, ensuring scalability, cost-efficiency, and observability.
- Integrate AWS Bedrock, Sage Maker, or OpenAI APIs with internal data systems and applications.
- Set up monitoring, tracing, and model observability using AWS Cloud Watch, X-Ray, or third-party LLMOps tools.
- Collaborate with ML engineers, data scientists, and architects to take GenAI prototypes to production-ready deployments.

Required

Skills & Experience:

- 6–10 years of total experience in Data Engineering with strong AWS background.
- Proficiency in Pyspark with hands-on production grade experience.
- Hands-on experience with GenAI solutions in real-world environments (not just demos or PoCs).
- Working knowledge of Agentic AI frameworks (Lang Chain, Llama Index, Auto Gen, or similar).
- Good hands-on experience in Python
- Cloud experience is must have, AWS is preferred.

- Experience with RAG architecture, vector databases (Open Search, Pinecone, FAISS, Chroma, or Milvus), and embedding models.
- Understanding of LLMOps, prompt lifecycle management, and performance monitoring.
- Practical experience deploying workloads on AWS ECS/EKS, setting up CI/CD pipelines, and managing runtime performance.
- Familiarity with IAM, VPC, Secrets Manager, and security best practices in cloud environments.

Nice to Have:

- Experience with AWS Bedrock for model hosting or Sage Maker for fine-tuning and evaluation.
- Exposure to multi-agent architectures and autonomous task orchestration.
- Contributions to open-source GenAI projects or internal AI platform initiatives.

For Quick Response
- Interested Candidates can directly share their resume along with the details like Notice Period, Current CTC and Expected CTC at
Note that applications are not being accepted from your jurisdiction for this job currently via this jobsite. Candidate preferences are the decision of the Employer or Recruiting Agent, and are controlled by them alone.
To Search, View & Apply for jobs on this site that accept applications from your location or country, tap here to make a Search:
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary