More jobs:
Intermediate Senior Data Engineer; Databricks
Job in
Centurion, 0014, South Africa
Listed on 2026-01-07
Listing for:
AICA Consultancy
Full Time
position Listed on 2026-01-07
Job specializations:
-
IT/Tech
Data Engineer, Data Science Manager, Cloud Computing, Big Data
Job Description & How to Apply Below
Overview
We are looking for a Data Engineer who is certified in Databricks (required) to join our team. In this role you will be designing, developing, and optimizing scalable data pipelines and workflows on Databricks. The engineer will work closely with stakeholders to make certain data reliability, performance, and alignment with business requirements.
Responsibilities Data Pipeline Development- Building efficient ETL / ELT pipelines using Databricks and Delta Lake for structured, semi-structured, and unstructured data.
- Transforming raw data into consumable datasets for analytics and machine learning.
- Improving performance by implementing best practices like partitioning, caching, and Delta Lake optimizations.
- Resolving bottlenecks and ensuring scalability.
- Integrating data from various sources such as APIs, databases, and cloud storage systems (e.g., AWS S3, Azure Data Lake).
- Designing and deploying real-time data streaming solutions using Databricks Structured Streaming.
- Implementing data validation, schema enforcement, and monitoring to ensure high-quality data delivery.
- Using Unity Cat Log to manage metadata, access permissions, and data lineage.
- Collaborating with data analysts, data scientists, and other stakeholders to meet business needs.
- Documenting pipelines, workflows, and technical solutions.
- Fully functional and documented data pipelines.
- Optimized and scalable data workflows on Databricks.
- Real-time streaming solutions integrated with downstream systems.
- Detailed documentation for implemented solutions and best practices.
- Proficiency in Databricks(certified), Spark, and Delta Lake.
- Strong experience with Python, SQL, and ETL / ELT development.
- Familiarity with real-time data processing and streaming.
- Knowledge of cloud platforms (e.g., AWS, Azure, GCP).
- Experience with data governance and tools like Unity Cat Log.
- Access to necessary datasets and cloud infrastructure will be provided.
- Timely input and feedback from stakeholders.
- Data pipelines deliver accurate and consistent data.
- Workflows meet performance benchmarks.
- Real-time streaming solutions operate with minimal latency.
- Stakeholders are satisfied with the quality and usability of the solutions.
Position Requirements
10+ Years
work experience
Note that applications are not being accepted from your jurisdiction for this job currently via this jobsite. Candidate preferences are the decision of the Employer or Recruiting Agent, and are controlled by them alone.
To Search, View & Apply for jobs on this site that accept applications from your location or country, tap here to make a Search:
To Search, View & Apply for jobs on this site that accept applications from your location or country, tap here to make a Search:
Search for further Jobs Here:
×