×
Register Here to Apply for Jobs or Post Jobs. X

Principal Engineer Machine Learning; MLOps DLP Detection

Job in Santa Clara, Santa Clara County, California, 95053, USA
Listing for: PaloAlto Networks
Full Time, Part Time position
Listed on 2026-01-07
Job specializations:
  • IT/Tech
    AI Engineer, Machine Learning/ ML Engineer, Cloud Computing
Job Description & How to Apply Below
Position: Principal Engineer Machine Learning (MLOps DLP Detection)

Company Description Our Mission

At Palo Alto Networks everything starts and ends with our mission:

Being the cybersecurity partner of choice, protecting our digital way of life.
Our vision is a world where each day is safer and more secure than the one before. We are a company built on the foundation of challenging and disrupting the way things are done, and we’re looking for innovators who are as committed to shaping the future of cybersecurity as we are.

Who We Are

We believe collaboration thrives in person. That’s why most of our teams work from the office full time, with flexibility when it’s needed. This model supports real-time problem-solving, stronger relationships, and the kind of precision that drives great outcomes.

Job Description Your Career

We are looking for a Principal MLOps Engineer to lead the design, development, and operation of production-grade machine learning infrastructure  this role, you will architect robust pipelines, deploy and monitor ML models, and ensure reliability, reproducibility, and governance across our AI/ML ecosystem. You will work at the intersection of ML, Dev Ops, and cloud systems, enabling our teams to accelerate experimentation while ensuring secure, efficient, and compliant deployments.

This

role is located at our dynamic Santa Clara California headquarters campus, and in office 3 days a week. Not a remote role. Your Impact
  • End-to-End ML Architecture and Delivery Ownership:
    Architect, design, and lead the implementation of the entire ML lifecycle. This includes ML model development and deployment workflows that seamlessly transition models from initial experimentation/development to complex cloud and hybrid production environments.
  • Operationalize Models at Scale:
    Develop and maintain highly automated, resilient systems that enable the continuous training, rigorous testing, deployment, real‑time monitoring, and robust rollback of machine learning models in production, ensuring performance meets massive scale demands.
  • Ensure Reliability and Governance:
    Establish and enforce state‑of‑the‑art practices for model versioning, reproducibility, auditing, lineage tracking, and compliance across the entire model inventory.
  • Drive Advanced Observability & Monitoring: Develop comprehensive, real‑time monitoring, alerting, and logging solutions focused on deep operational health,
    model performance analysis (e.g., drift detection), and business metric impact.
  • Champion Automation & Efficiency: Act as the primary driver for efficiency, pioneering best practices in Infrastructure‑as‑Code (IaC), sophisticated container orchestration, and continuous delivery (CD) to reduce operational toil.
  • Collaborate and Lead Cross‑Functionally: Partner closely Security Teams, and Product Engineering to define requirements and deliver robust, secure, and production‑ready AI systems.
  • Lead MLOps Innovation: Continuously evaluate, prototype, and introduce cutting‑edge tools, frameworks, and practices that fundamentally elevate the scalability, reliability, and security posture of our production ML operations.
  • Optimize Infrastructure & Cost: Strategically manage and optimize ML infrastructure resources to drive down operational costs, improve efficiency, and reduce model bootstrapping times.
Qualifications Your Experience
  • 8 years of software/Dev Ops/ML engineering experience, with at least 3 years focused specifically on advanced MLOps, ML Platform, or production ML infrastructure and 5 yeas of experience building ML Models
  • Deep expertise in building scalable, production‑grade systems using strong programming skills (Python, Go, or Java).
  • Expertise in leveraging cloud platforms (AWS, Google Cloud Platform, Azure) and container orchestration (Kubernetes, Docker) for ML workloads.
  • Proven hands‑on experience in the ML Infrastructure lifecycle, including:
  • Model Serving: (Tensor Flow Serving, Torch Serve, Triton Inference Server/TIS).
  • Workflow Orchestration: (Airflow, Kubeflow, MLflow, Ray, Vertex AI, Sage Maker).
  • Mandatory Experience with Advanced Inferencing Techniques: Demonstrable ability to utilize advanced hardware/software acceleration and optimization techniques, such as Tensor

    RT (TRT), Triton…
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary