×
Register Here to Apply for Jobs or Post Jobs. X

Apache Iceberg Engineer

Job in Sunnyvale, Santa Clara County, California, 94087, USA
Listing for: Smart IT Frame LLC
Full Time position
Listed on 2025-12-01
Job specializations:
  • IT/Tech
    Data Engineer, Cloud Computing, Big Data, Data Science Manager
Job Description & How to Apply Below

We are looking for an experienced Apache Iceberg Engineer to design, develop, and optimize large-scale data lakehouse solutions leveraging Apache Iceberg. The ideal candidate will have expertise in big data processing frameworks (Apache Spark, Flink, Presto, Trino, Hive) and cloud-based data platforms like AWS S3, Google Cloud Storage, or Azure Data Lake Storage. You will work closely with data engineers, data scientists, and Dev Ops teams to ensure efficient, scalable, and reliable data architecture.

Key Responsibilities:

  • Design, implement, and optimize Iceberg-based data lake architectures for large-scale datasets.
  • Develop data ingestion, transformation, and query optimization pipelines using Spark, Flink, or Presto/Trino.
  • Ensure ACID compliance, schema evolution, and partition evolution in Iceberg tables.
  • Implement time travel, versioning, and snapshot management for historical data analysis.
  • Optimize metadata management and query performance in Iceberg-based data lakes.
  • Integrate Apache Iceberg with cloud storage solutions (AWS S3, GCS, ADLS) and data warehouses.
  • Implement best practices for data governance, access control, and security within an Iceberg-based environment.
  • Troubleshoot performance issues, metadata inefficiencies, and schema inconsistencies in Iceberg tables.
  • Collaborate with Dev Ops, ML engineers, and BI teams to enable smooth data workflows.

Required Qualifications:

  • Bachelor's or Master's degree in Computer Science, Data Engineering, or a related field.
  • 3+ years of experience in Big Data, Data Engineering, or Cloud Data Warehousing.
  • Hands-on experience with Apache Iceberg in a production environment.
  • Strong expertise in Apache Spark, Flink, Trino, Presto, or Hive for big data processing.
  • Proficiency in SQL and distributed query engines.
  • Experience working with cloud storage solutions (AWS S3, GCS, ADLS).
  • Knowledge of data lakehouse architectures and modern data management principles.
  • Familiarity with schema evolution, ACID transactions, and partitioning techniques.
  • Experience with Python, Scala, or Java for data processing.

Preferred Qualifications:

  • Experience in real-time data processing using Flink or Kafka.
  • Understanding of data governance, access control, and compliance frameworks.
  • Knowledge of other data lake frameworks like Delta Lake (Databricks) or Apache Hudi.
  • Hands-on experience with Terraform, Kubernetes, or Airflow for data pipeline automation.
Seniority level

Mid-Senior level

Employment type

Full-time

Job function

Other

Industries

Software Development and IT Services and IT Consulting

#J-18808-Ljbffr
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary