×
Register Here to Apply for Jobs or Post Jobs. X
More jobs:

Senior Data Engineer

Job in Coos Bay, Coos County, Oregon, 97458, USA
Listing for: Amwell
Full Time position
Listed on 2026-01-02
Job specializations:
  • IT/Tech
    Data Engineer
Job Description & How to Apply Below

Company Description

At Amwell, we’re transforming healthcare for all—powered by technology and inspired by people. Here, your ideas don’t just matter—they drive real change, improving lives on a global scale.

We marry technology and innovation with clinical excellence to provide trusted solutions that solve the healthcare industry’s biggest pain points and are on a mission to enable greater access to more convenient, affordable, and effective care.

We do this through our technology-enabled care platform that is designed to help our clients achieve their digital care ambitions – today and in the future. We offer programs spanning the full care continuum
, including urgent, acute and specialty care, behavioral health, and services for the treatment of chronic conditions such as heart and cardiometabolic diseases. Programs are powered by Amwell as well as our growing partner network.

For almost two decades, Amwell has proudly served some of the largest and most sophisticated healthcare organizations in the U.S. and worldwide. Our team is passionate about technology’s role in transforming care delivery and making it more equitable, accessible, efficient, cost-effective and navigable for all.

Brief Overview

The Senior Data Engineer should have experience in end-to-end implementation of data-warehousing projects, he will manage, utilize, move, and transform data from our source system and applications data to the cloud to create reports for senior management and internal users. working both independently on assigned projects, and collaboratively with other team members. Build various ETL pipelines among the various tools in play to surface data for consumption by our Reporting tool.

Prioritize competing requests from internal and external stakeholders in addition to keeping the reporting infrastructure on par with new product functionality and release cycles. Become a Subject Matter Expert in data classification within the platform and utilize your expertise to identify the most efficient path to deliver data from Source to target as needed.

Core Responsibilities
  • Design and write excellent, fully tested code to ETL/ELT data pipelines and stream on a cloud platform.
  • Good communication skills, as well as the ability to work effectively across internal and external organizations and virtual teams.
  • Implement product features and refine specifications with our product manager and product owners.
  • Stay familiar with industry changes, especially in the areas of cloud data and analytics technologies.
  • Ability to work on multiple areas like Data pipeline ETL, Data modelling design, writing complex SQL queries etc. and a good understanding of BI/DWH principles
  • Capable of planning and executing both short-term and long-term goals individually and with the team.
  • Understanding of the SDLC (Software Development life cycle) and Knowledge of Scrum, Agile
  • On-call rotation and Production Jobs monitoring.
Qualifications
  • 5+ years of development experience building data pipelines.
  • Bachelor’s Degree or equivalent experience is required. Preferred in Computer Science or related degree
  • Minimum of 5 years of experience in architecture of modern data Warehousing platforms using technologies such as Big Data and Cloud, Kafka experience.
  • Cloud experience - any cloud, preferably Bigquery, data flow, pub-sub, data fusion
  • Migration experience, Utilizing GCP to move data from on-prem servers to the cloud
  • Good Python development for data transfers and extractions (ELT or ETL)
  • Experience developing and deploying ETL solutions like Informatica or similar tools
  • Experience working within an agile development process (Scrum, Kanban, etc)
  • Familiarity with CI/CD concepts
  • Demonstrated proficiency in creating technical documentation
  • Modern concepts (how new-gen DB is implemented – like how BQ/Redshift works?).
  • Airflow, Dag development experience
  • Informatica or any ETL tool previous experience
  • Ability and experience in BI and Data Analysis, end-to-end development in data platform environments.
  • Modern concepts (how new-gen DB is implemented – like how BQ/Redshift works?).
  • Fix things before they break
  • write excellent, fully tested code to build ETL /ELT data pipelines on…
Position Requirements
10+ Years work experience
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary