Data Product Engineer APIFocused FullStack Data Engineer
Listed on 2026-01-05
-
IT/Tech
Data Engineer, Data Science Manager
Job Description
Founded in 2002, Source Allies is an agile, team-based technical consultancy focused on building industry-defining data and cloud-native solutions. Our teammates partner closely with clients to design, build, and evolve data products that turn raw data into actionable insights, delivered through consumer-ready datasets and API-first interfaces.
We’re seeking a Data Product Engineer who blends software engineering and data engineering and can navigate ambiguity with a “data-as-a-product” mindset. You’ll help build and run production-grade data products, often surfaced through APIs (including federated Graph
QL), with data contracts, observability, and measurable outcomes.
Build data products with real consumers and outcomes
- Collaborate with product managers and stakeholders to translate complex business needs and questions into insights delivered through data products.
- Define and iterate on product success metrics (e.g., adoption, time-to-insight, reliability, cost-to-serve), and improve the product based on usage and feedback.
- Treat data assets and APIs as products with lifecycle ownership: launch, adoption, improvement, and retirement.
Engineer API-first data solutions
- Design and develop data product APIs using Java, AWS, and modern CI/CD practices.
- Contribute to or build federated Graph
QL APIs (potentially using Apollo Graph
OS) and/or REST interfaces, with strong attention to contracts, versioning, and developer experience. - Build production-ready services with observability (e.g., Datadog) and infrastructure-as-code (e.g., Terraform).
Build reliable pipelines and analytics foundations
- Evaluate and compare data sources to determine optimal solutions, using quantitative analysis and tradeoff reasoning.
- Explore data sources to uncover attributes, understand lineage/data flows, detect trends, and validate feasibility.
- Build and maintain scalable, testable pipelines using modern Python engineering practices (modules, tests, CI/CD) as opposed to standalone Databricks notebooks.
- Perform analysis and pipeline development using Databricks, PySpark, and SQL.
Operate with quality and craft
- Apply Agile and XP practices such as TDD, pairing, small batch delivery, and frequent demos/feedback loops.
- Establish and maintain expectations for data quality, freshness, latency, and reliability through testing, monitoring, and automation.
- Collaborate across engineering, product, and analytics teams to ensure data products are discoverable, trusted, and reusable.
Core experience
- 3+ years of industry experience building enterprise software, data platforms, or data-driven products.
- Strong foundation in software engineering and data engineering: you can design systems, write clean code, and ship production services.
- Proficiency in Python, SQL, and Java (or equivalent OO language with willingness to work in Java).
- Cloud experience, preferably AWS, including building and operating production services.
Data product mindset
- You think beyond pipelines: you build consumer-ready data assets and data solutions with clear contracts, documentation, and usability.
- Comfortable navigating ambiguity and working independently to deliver an answer, a fix, or an enhancement end-to-end.
- You can translate business questions into analytical approaches, validate assumptions, and deliver outcomes.
Technical strengths aligned to this role
- Experience designing and building APIs that deliver data/insights (Graph
QL and/or REST), including schema design and backwards compatibility. - Experience building pipelines with Databricks, PySpark, and SQL.
- Familiar with CI/CD (e.g., Git Hub Actions) and infrastructure tooling (e.g., Terraform).
- Observability experience (e.g., Datadog) across services and pipelines.
- Familiarity with data quality practices (tests/validation, monitoring, freshness checks, lineage awareness).
- Experience with federated Graph
QL and/or Apollo Graph
OS. - Experience with Elastic / Elasticsearch.
- Experience with modern data warehousing/lakehouse tools (Snowflake, Big Query, Redshift, Athena, etc.).
- Exposure to BI/semantic layers and tools (Power BI, Tableau, Looker), especially when paired with API-first delivery patterns.
- Competitive salary and benefits
- Health, Dental, Vision, Life, Disability, Maternity & Paternity Leave benefits
- 401(k) match with immediate vesting
- Ongoing training and learning opportunities because continuous improvement matters
Join Source Allies and help clients turn data into a competitive advantage. In our hybrid work model, all teammates have some level of onsite requirement, balancing flexibility with meaningful in-person collaboration.
Company DescriptionSource Allies is a local technical consultancy that values collaboration, craftsmanship, and an ownership mentality. We provide a home to our teammates where we encourage each other to grow not only technically, but as peers & consultants. Find your Mastery Path at Source Allies! See more at
#J-18808-Ljbffr(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).