×
Register Here to Apply for Jobs or Post Jobs. X
More jobs:

Data Architect

Job in San Rafael, Marin County, California, 94911, USA
Listing for: Incedo Inc.
Full Time position
Listed on 2026-01-01
Job specializations:
  • IT/Tech
    Data Engineer
Job Description & How to Apply Below
Senior Talent Acquisition Specialist at Incedo - Hiring for Life Sciences

Role Overview

We are seeking a Data bricks Data Architect to support the design, implementation, and optimization of cloud-native data platforms built on the Data bricks Lakehouse Architecture. This is a hands-on, engineering-driven role requiring deep experience with Apache Spark, Delta Lake, and scalable data pipeline development, combined with early-stage architectural responsibilities.

The role involves close onsite collaboration with client stakeholders, translating analytical and operational requirements into robust, high-performance data architectures, while adhering to best practices for data modeling, governance, reliability, and cost efficiency.

Key Responsibilities

• Design, develop, and maintain batch and near-real-time data pipelines using Databricks, PySpark, and Spark SQL

• Implement Medallion (Bronze/Silver/Gold) Lakehouse architectures, ensuring proper data quality, lineage, and transformation logic across layers

• Build and manage Delta Lake tables, including schema evolution, ACID transactions, time travel, and optimized data layouts

• Apply performance optimization techniques such as partitioning strategies, Z-Ordering, caching, broadcast joins, and Spark execution tuning

• Support dimensional and analytical data modeling for downstream consumption by BI tools and analytics applications

• Assist in defining data ingestion patterns (batch, incremental loads, CDC, and streaming where applicable)

• Troubleshoot and resolve pipeline failures, data quality issues, and Spark job performance bottlenecks.

Nice-to-Have Skills

• Exposure to Data bricks Unity Catalog, data governance, and access control models

• Experience with Data bricks Workflows, Apache Airflow, or Azure Data Factory for orchestration

• Familiarity with streaming frameworks (Spark Structured Streaming, Kafka) and/or CDC patterns

• Understanding of data quality frameworks, validation checks, and observability concepts

• Experience integrating Data bricks with BI tools such as Power BI, Tableau, or Looker

• Awareness of cost optimization strategies in cloud-based data platforms

#JLjbffr
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary