×
Register Here to Apply for Jobs or Post Jobs. X

Data engineer

Job in Riyadh, Riyadh Region, Saudi Arabia
Listing for: Master Works
Full Time position
Listed on 2026-03-10
Job specializations:
  • IT/Tech
    Data Engineer, Data Warehousing
Salary/Wage Range or Industry Benchmark: 150000 - 200000 SAR Yearly SAR 150000.00 200000.00 YEAR
Job Description & How to Apply Below

Master Works is hiring an experienced Data Engineer (5+ years) in Riyadh to design and optimize large-scale real-time and batch data pipelines within the Telecom domain.

  • Design, develop, and maintain real-time and batch data pipelines leveraging Kafka, Spark, and Hadoop components.
  • Must have understanding of Teradata CLDM, should know how to create new Data Model or modify existing data model based on business requirement
  • Collaborate with business analysts and data architects to translate business requirements into robust data models and ETL frameworks.
  • Apply Relational and Dimensional modeling techniques to design databases and ensure data is organized effectively for both operational and analytical purposes.
  • Write, debug, and optimize SQL and Stored Procedures to ensure efficient data processing.
  • Work closely with BI, Data Science, and Campaign teams to ensure seamless data availability for analytics
  • Work closely with Data Architects, Analysts, and Business Stakeholders to translate business requirements into database solutions.
  • Ensure all database design and code is well-documented and follows best practices for performance and maintainability.
  • Involves designing fact and dimension tables for reporting and analytics purposes, often in a star or snowflake schema.
  • Develop and maintain technical documentation (data flow diagrams, source-to-target mappings, architecture documents).
  • Ensure that the Data Dictionary is always up-to-date, capturing all changes to the database schema, including newly created or modified tables, columns, views,
  • Perform data quality checks, validation, and ensure end-to-end data accuracy and lineage.
  • Support and troubleshoot real-time streaming jobs and ensure high availability of data pipelines
Have-Must
  • Strong expertise in real-time data integration using Kafka, Spark Streaming, or Dataflow.
  • Hands-on experience with Hadoop ecosystem components (HDFS, Hive, Sqoop, Spark etc.).
  • Strong Data Modeling concepts including FSLDM, CLDM, and Dimensional / Data Vault modeling.
  • Deep understanding of Telecom domain (BSS/OSS, CDR, usage, revenue, and campaign data).
  • Experience building and optimizing ETL pipelines and data ingestion frameworks for structured and unstructured data.
  • Proficiency in SQL and distributed data processing using Hive, Spark SQL, or PySpark.
  • Good understanding of data governance, data quality, and lineage frameworks.
  • Strong analytical and problem-solving skills.
  • Excellent communication and collaboration skills with cross-functional teams.
  • Experience in working on data vartulization tools like (Tibco. Trino etc)
Good-to-Have
  • Familiarity with Data Catalogs, Metadata Management, and NDMO data governance standards.
  • Experience with Data Catalogue tools.
  • Familiarity with CI/CD pipelines, Git.
  • Knowledge of ETL orchestration tools like Airflow, Ni Fi
#J-18808-Ljbffr
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary