GCP Data engineer
Listed on 2026-02-17
-
IT/Tech
Data Engineer, Big Data, Database Administrator, Data Warehousing
Skill 1 – 8+ years of experience in Python, SQL, and potentially Scala/Java
Skill 2 – Big Data:
Expertise in Apache Spark (Spark SQL, Data Frames, Streaming).
Skill 3- 4+ Years in GCP
Data Engineer with Spark & Streaming skills builds real-time, scalable data pipelines using tools like Spark, Kafka, and cloud services (GCP) to ingest, transform, and deliver data for analytics and ML.
Responsibilities:Design, develop, and maintain ETL/ELT data pipelines for batch and real-time data ingestion, transformation, and loading using Spark (PySpark/Scala) and streaming technologies (Kafka, Flink).
Build and optimize scalable data architectures, including data lakes, data warehouses (Big Query), and streaming platforms.
Performance Tuning:
Optimize Spark jobs, SQL queries, and data processing workflows for speed, efficiency, and cost-effectiveness
Data Quality:
Implement data quality checks, monitoring, and alerting systems to ensure data accuracy and consistency.
Qualifications:
Programming:
Strong proficiency in Python, SQL, and potentially Scala/Java.
Big Data:
Expertise in Apache Spark (Spark SQL, Data Frames, Streaming).
Streaming:
Experience with messaging queues like Apache Kafka, or Pub/Sub.
Cloud:
Familiarity with GCP, Azure data services.
Databases:
Knowledge of data warehousing (Snowflake, Redshift) and No
SQL databases.
Tools:
Experience with Airflow, Databricks, Docker, Kubernetes is a plus.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).