×
Register Here to Apply for Jobs or Post Jobs. X

Full Stack Machine Learning Engineer; Datacentre AI Engineering KSA

Job in Riyadh, Riyadh Region, Saudi Arabia
Listing for: Qualcomm Technologies, Inc
Full Time position
Listed on 2025-12-06
Job specializations:
  • IT/Tech
    AI Engineer, Machine Learning/ ML Engineer
Salary/Wage Range or Industry Benchmark: 300000 - 400000 SAR Yearly SAR 300000.00 400000.00 YEAR
Job Description & How to Apply Below
Position: Full Stack Machine Learning Engineer (Data centre AI Engineering KSA

Company: Qualcomm Middle East Information Technology Company LLC

Job Area: Engineering Group, Engineering Group >
Software Engineering

About Us

Qualcomm is enabling a world where everyone and everything can be intelligently connected. You interact with products and technologies made possible by Qualcomm every day, including 5G-enabled smartphones that double as pro-level cameras and gaming devices, smarter vehicles and cities, and the technology behind the smart, connected factories that manufactured your latest purchase. Qualcomm 5G and AI innovations are the power behind the connected intelligent edge.

You’ll find our technologies behind and inside the innovations that deliver significant value across multiple industries and to billions of people every day.

Job Overview

We are seeking a Full-Stack Machine Learning Engineer to join our team, bridging AI solutions development with AI platform engineering for Qualcomm’s AI Inference Suite and rack-scale data center deployments. This role involves designing, delivering, and supporting end-to-end AI services, agentic workflows, and fine-tuning pipelines, while enabling lifecycle automation, orchestration, and observability for large-scale data center environments.

You will bring a strong combination of full-stack engineering expertise, machine learning proficiency, and infrastructure knowledge to build robust, scalable AI systems.

Key Responsibilities
  • API Development & Optimization:
    Build and optimize API serving layers for AI inference workloads, ensuring model and hardware efficiency.
  • Agentic Workflows & RAG Pipelines:
    Develop intelligent agents and retrieval-augmented generation workflows using frameworks such as Lang Chain and crew.ai.
  • Model Lifecycle Management:
    Implement production-grade bring-your-own-model and fine-tuning flows, including dataset ingestion, orchestration, evaluation, and deployment.
  • LLM Runtime Integration:
    Work with various LLM runtimes (e.g., vLLM, Dynamo, llm-d) and leverage inference optimization techniques.
  • SDK & Tooling Contributions:
    Contribute to AI Inference Suite SDKs (Python/Type Script/Java/Rust), CLI tools, and reference applications.
  • Cluster Management:
    Design and maintain AI cluster management software for provisioning, orchestration, and monitoring.
  • Telemetry & Observability:
    Integrate out-of-band management via Redfish/IPMI and in-band telemetry using Prometheus/Open Telemetry.
  • Infrastructure-as-Code:
    Develop workflows using MAAS, Terraform, and Ansible for bare-metal and containerized deployments.
  • Kubernetes Orchestration:
    Enable Kubernetes/Helm-based orchestration for inference clusters and multi-tenancy.
  • Monitoring & Dashboards:
    Build dashboards for rack health, inventory, and SLA compliance.
  • Continuous Innovation:
    Stay current with GenAI trends, rack-scale AI orchestration, and data center best practices.
Minimum Qualifications
  • Bachelor’s degree in Computer Science, Engineering, or related field.
  • 5+ years of software engineering experience; 3+ years in ML or HPC environments.
  • Strong programming skills in Python, Rust/Go, and Type Script, with solid software development fundamentals.
  • Deep understanding of data structures and algorithms in distributed systems and high-performance computing contexts.
  • Hands‑on experience with Kubernetes, Helm, Prometheus/Open Telemetry, and Ansible/Terraform.
  • Practical experience with LLM runtimes, agent frameworks, and rack‑scale orchestration.
Preferred Qualifications
  • Master’s degree in Computer Science, Machine Learning, or related field.
  • Experience building inference and fine‑tuning pipelines, as well as agentic workflows.
  • Knowledge of data centre resource lifecycle management, out-of-band protocols (Redfish/IPMI), and MAAS/Open Stack.
  • Exposure to scale‑up data centre networking technologies (RoCE/RDMA/NVLink).
  • Contributions to inference and GenAI model performance optimization.
What's on Offer

Apart from working with great people, we offer the below:

  • Salary including housing & transport allowance
  • Stock (RSU's) and performance related bonus
  • 16 weeks fully paid Maternity Leave
  • 6 weeks fully paid Paternity Leave
  • Employee stock purchase scheme
  • Child Education Allowance
  • Relocation and immigration support…
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary