×
Register Here to Apply for Jobs or Post Jobs. X

Azure Engineer

Job in Aliso Viejo, Orange County, California, 92656, USA
Listing for: Saransh Inc
Full Time position
Listed on 2026-03-04
Job specializations:
  • IT/Tech
    Data Engineer, Cloud Computing
Salary/Wage Range or Industry Benchmark: 100000 - 125000 USD Yearly USD 100000.00 125000.00 YEAR
Job Description & How to Apply Below
Position: Azure Engineer - Full Time

Role: Data Engineer

Location: Remote (Candidate should be comfortable following PST Time Zone)

Job Type: Full Time

Active Linked In  a must and must match with the resume.

Must Have Skills
  • Azure Databricks, Apache Spark, Snowflake, Data Governance, Unity Catalog, dBT, Microsoft Purview, Kafka, SQL, Python
Experience And Qualifications
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
  • 10+ years of experience in data engineering / data architecture, with deep hands‑on expertise in Microsoft Azure.
  • Expert‑level experience with Azure Databricks and Apache Spark (PySpark, Spark SQL).
  • Delta Lake, Unity Catalog, Job orchestration and performance tuning.
  • Strong experience with Snowflake on Azure, including:
    • Schema and data model design
    • Performance and cost optimization
    • RBAC, Streams, and Tasks
  • DBT (Core or Cloud) expertise:
    • Project structure and best practices
    • Tests, exposures, macros
  • Strong fundamentals in SQL and Python.
  • Proven experience building ETL/ELT pipelines using Azure Data Factory (ADF) and Azure Synapse / Microsoft Fabric pipelines.
  • Familiarity with streaming platforms such as Event Hubs and/or Kafka.
  • Solid understanding of Lakehouse and data warehousing architectures, dimensional modeling, medallion architecture patterns, and data quality frameworks.
  • Strong knowledge of security and governance, including Microsoft Purview (catalog, lineage), data masking and PII handling, managed identities, and private networking.
  • Excellent communication and documentation skills, with the ability to create architecture and design documents and present to both technical and business stakeholders.
  • Consulting / agency experience, with the ability to lead multiple concurrent projects.
  • Willingness to travel as required.
Nice To Have
  • Experience with Azure ML, feature stores, or serving ML pipelines from Databricks or Snowflake.
  • Infrastructure as Code (IaC) using Terraform or Bicep.
  • Containerization experience with Docker.
  • Observability and data quality tooling, such as Great Expectations / Delta Expectations, Databricks Quality Flows, Monte Carlo, Datadog.
Preferred Certifications
  • Databricks Certified Data Engineer Professional or Databricks Architect.
  • Microsoft Azure: DP-203 – Azure Data Engineer, Azure Solutions Architect Expert, Azure Security Engineer Associate.
  • Snowflake:
    Snow Pro Core, Snow Pro Advanced.
Job Description / Responsibilities
  • Partner with client stakeholders to translate business objectives into scalable Azure data architectures and delivery roadmaps.
  • Architect and implement lakehouse and data warehouse solutions using Azure Databricks (Delta Lake, Unity Catalog), Databricks Bundle, Snowflake on Azure, dbt (Core / Cloud), ADLS Gen2, and medallion architecture patterns.
  • Design and build robust ingestion and transformation pipelines using Azure Data Factory (ADF) and Microsoft Fabric / Synapse pipelines.
  • Orchestrate ELT workflows using Databricks Jobs, Delta Live Tables, and dbt models.
  • Establish best practices for performance and cost optimization, including cluster sizing and autoscaling, Photon and SQL Warehouses, Snowflake virtual warehouses, caching, file layout optimization, and Z‑ordering.
  • Drive observability and monitoring using Azure Monitor, Log Analytics, and Databricks metrics.
  • Implement robust data governance and security using Azure Authentication, Unity Catalog, RBAC / ABAC, managed identities, private endpoints and VNet injection, and Azure Key Vault–backed secrets.
  • Lead code and design reviews, setting standards for PySpark, SQL, and dbt, unit and integration testing, and data quality checks (expectations and constraints).
  • Implement CI/CD pipelines using Azure Dev Ops or Git Hub Actions.
  • Guide multi‑domain programs (Healthcare, Retail, BFSI, or similar).
  • Mentor data engineers and ensure high‑quality, production‑ready deliverables.
  • Evangelize solutions through clear documentation, reference architectures, and executive and stakeholder presentations.
#J-18808-Ljbffr
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary