Senior Data Tech Lead/Architect
Listed on 2026-02-14
-
IT/Tech
Data Engineer, Cloud Computing
Title:
Senior Data Tech Lead/Architect – AWS | Kafka | | Glue Streaming; API consumption
Location:
Dallas, Texas (Preferred) Or Florham Park, New Jersey , NJ Full time
About the Role:
We are seeking a Senior Technical Lead
/
Architect with strong expertise in AWS-based streaming data pipelines
,
Apache Kafka (MSK),
AWS Glue
, Flink
and
Py Spark , to help solution, design and implement a scalable data ingestion, data validation, data enrichment and reconciliation processing, and event logs, data observability, operational KPI trackingframework.
You will play a key role in solutioning and building out the, event driven capabilities with control gates in place to measure, track, and improve the operational SLAs and drive the
data quality and reconciliation workflows
for a high-impact data platform serving financial applications
- Provide technical solution discovery effort on any new capabilities or new functionality.
- Assist PO with technical user stories to ensure healthy backlog features
- Lead the development of
real-time data pipelines
using
AWS DMS
,
MSK
, Kafka or Glue Streaming andfor CDC ingestion from multiple SQL Server sources (RDS/on-prem). - Build and optimize
streaming and batch data pipelines
using
AWS Glue (PySpark)to validate, transform, and normalize data to
Iceberg and DynamoDB
. - Define and enforce
data quality, lineage, and reconciliation logic
with support for both streaming and batch use cases. - Integrate with
S3 Bronze/Silver layers
and implement efficient
schema evolution and partitioning strategies
using Iceberg. - Collaborate with architects, analysts, and downstream application teams to design
API and file-based egress
layers. - Implement
monitoring, logging
, and
event-based alerting
using Cloud Watch, SNS, and Event Bridge. - Mentor junior developers and enforce best practices for modular, secure, and scalable data pipeline development.
- 6+ years of hands‑on expert level data engineering experience in
cloud‑based environments
(AWS preferred) with event driven implementation - Strong experience with
Apache Kafka / AWS MSK
including topic design, partitioning, and Kafka Connect/Debezium - Proficiency in
AWS Glue (PySpark)and for both batch and streaming ETL - Working knowledge of
AWS DMS
,
S3
,
Lake Formation
,
DynamoDB
, and
Iceberg - Solid grasp of
schema evolution
,
CDC patterns
, and
data reconciliation frameworks - Experience with infrastructure‑as‑code (CDK/Terraform) and Dev Ops practices (CI/CD, Git)
- Familiarity with
PostgreSQL
,
Athena
,
Step Functions
, and
Lambda - Knowledge of
event‑driven architectures
and
microservices‑based data APIs (This should be the required skills) - Familiar with data lake concept
- Familiar with Financial or Financial Wealth management data domain
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).