Senior Data Architect
Listed on 2026-02-16
-
IT/Tech
Data Engineer, Cloud Computing
Senior Data Architect
Boston, MA (Hybrid)
4+ months contract (Possible extension)
OSA serves as the chief accountability entity for the Massachusetts state government and its residents. OSA conducts audits of state entities and contractors to assess their performance and recommend improvements to enhance the effectiveness of government operations. In addition to ensuring that tax dollars are spent wisely, audits, reports, and investigations have also improved the performance of state governments. OSA has offices in Boston, Marlborough, Chicopee, and Brockton.
DetailsClient is seeking a Senior Data Architect to design, implement, and operationalize a data lakehouse architecture within OSA's Azure Government Community Cloud (GCC) tenancies. You will establish the platform, governance, ingestion, transformation, and consumption layers, and enable our data engineers for ongoing operations. You will be required to work with developers, cybersecurity engineers, and other OSA staff to review the environment and implement best practices.
Key Responsibilities- Architecture & Platform
- Define and establish the target reference architecture for an enterprise lakehouse, with Azure Databricks as the foundational platform.
- Design and implement ADLS Gen2 structure and Medallion (Bronze/Silver/Gold) model using Delta Lake.
- Establish Unity Catalog, the metastore, and semantic layers.
- Governance & Security
- Implement Microsoft Purview for catalog, lineage, and metadata management.
- Design, align, and integrate data security policies (RBAC, RLS/CLS, masking) and PII handling.
- Integrate Key Vault, private endpoints, and network security baselines.
- Data Engineering Enablement
- Build batch/streaming ingestion pipelines (ADF/Synapse/Databricks Workflows, Event Hubs/Kafka).
- Implement change data capture (CDC) patterns and schema evolution handling with quarantine controls.
- Support automation of data reliability assessment.
- Create reusable transformation libraries, standards, and pipelines for Bronze→Silver→Gold.
- Operations and Reliability
- Establish continuous integration/continuous deployment (CI/CD) pipelines, along with infrastructure-as-a-code orchestration (Terraform or Bicep), and segregated environment strategy (dev/test/prod).
- Define service level agreements (SLAs) and service level objectives (SLOs) for data reliability; configure monitoring & alerting.
- Deliver cost optimization, performance tuning (OPTIMIZE, Z-ORDER), Backup strategy, and Disaster Recovery strategy.
- Delivery and Handover
- Document architecture, patterns, and runbooks; conduct training for data engineers.
- Collaborate with security, networking, and BI teams; drive stakeholder alignment.
- 8-10 years in data architecture/engineering, 3-5 years building lake houses in Azure. Experience operating within an GCC environment is highly desirable.
- Hands‑on with Azure Databricks (Delta Lake, Unity Catalog, Jobs, SQL).
- Strong with ADLS Gen2, Azure Data Factory Pipelines, Event Hubs/Kafka, Purview, Key Vault.
- Expertise in Delta Lake, schema evolution, SCDs, CDC, streaming (Structured Streaming).
- Proficient in Python and SQL; experience with CI/CD and IaC (Terraform or Bicep).
- Security‑first mindset: RBAC, RLS/CLS, PII governance, private networking.
- Excellent communication, documentation, and stakeholder management.
- Expertise in engagement with business users to align business objectives with technical considerations within Azure.
- Microsoft Certified:
Azure Data Engineer Associate or Databricks Certified Data Engineer Professional. - Experience integrating Power BI semantic models with Delta tables.
- Familiarity with data quality frameworks (Great Expectations, Databricks expectations).
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).