Product Support Engineer - Hadoop Operations
Listed on 2025-12-13
-
Engineering
Data Engineer, Systems Engineer -
IT/Tech
Data Engineer, Systems Engineer
Location: California
We are seeking aStaff Product Support Engineer / Product Specialist - Hadoop Operations SME who will be responsible for designing, optimizing, migrating, and scaling Hadoop and Spark-based data processing systems. This role involves hands‑on experience with Hadoop and other core data operations, focusing on building resilient, high‑performance distributed data systems.
You will collaborate with customer engineering teams to deliver high‑throughput Hadoop, Nifi, and Spark applications and solve complex data challenges in migration, upgrades, reliability, and optimise post‑migration system performance.
This role requires flexibility to work in rotational shifts, based on team coverage needs and customer demand. Candidates should be comfortable supporting operations in a 24/7 environment and be willing to adjust their working hours accordingly.
We’re looking for someone who perform:- Design and optimise distributed Hadoop‑based applications, ensuring low‑latency, high‑throughput performance for big data workloads.
- Troubleshooting:
Provide expert‑level support for data or performance issues in Hadoop, Nifi, and Spark jobs and clusters. - Data Processing Expertise:
Work extensively with large‑scale data pipelines using Hadoop, Nifi, and Spark's core components. - Performance Tuning:
Conduct deep‑dive performance analysis, debugging, and optimisation of Nifi, Impala, and Spark jobs to reduce processing time and resource consumption. - Cluster Management:
Collaborate with Dev Ops and infrastructure teams to manage Nifi, Impala, and Spark clusters on platforms like Hadoop/YARN, Kubernetes, or cloud platforms (AWS EMR, GCP Dataproc, etc.). - Migration:
Provide dedicated support to ensure the stability and reliability of the new ODP Hadoop environment during and after migration. Promptly address evolving technical challenges and optimise system performance after migration to ODP.
- Master’s degree and Experience working with scripting languages (Scala, Python, Bash, Power Shell).
- Familiarity with virtual machine technologies and multi‑node environment (50+ nodes).
- Proficient with Linux, NFS, and Windows, including application installation, scripting, and working with the command line.
- Working knowledge of application, server, and network security management concepts. Certification on any of the leading Cloud providers (AWS, Azure, GCP ) and/or Kubernetes.
- Knowledge of databases like MySQL and Postgre
SQL. - Be involved with and work on other support‑related activities - Performing POC & assisting with Onboarding deployments of Acceldata & Hadoop distribution products.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).