Data Engineer
Job in
Houston, Harris County, Texas, 77246, USA
Listed on 2026-02-16
Listing for:
Comfort Systems USA, Inc.
Full Time
position Listed on 2026-02-16
Job specializations:
-
IT/Tech
Data Engineer, Data Analyst, Data Science Manager, Data Warehousing
Job Description & How to Apply Below
** The Data Engineer is responsible for designing, building, and maintaining the organization’s enterprise data pipelines, curated datasets, and analytics platform foundation. This role enables timely, accurate, and trusted decision-making by ensuring reliable, scalable, and governed data flows across the Enterprise Resource Planning (ERP) system, Product Lifecycle Management (PLM) system, and other integrated business applications.
The Data Engineer will engineer ingestion and curation from vendor-hosted enterprise data repositories and integration services, including Infor OS Data Lake and supported extraction methods (e.g., Data Fabric Objects API, Compass SQL API / Compass JDBC Driver, ION Data Lake Flows, and ETL Client for Data Lake). The role will enable enterprise analytics and reporting through Microsoft Fabric and Power BI, and will use Spark/PySpark, SQL, and Python to deliver performant, scalable transformations.
This role also supports the organization’s increasing use of AI-enabled analytics by preparing high-quality, well-governed datasets and following company requirements for responsible and compliant AI usage.
** KEY JOB RESPONSIBILITIES
**** Data Engineering & Architecture (Microsoft Fabric / Lakehouse)
*** Design and maintain a modern enterprise analytics foundation using
** Microsoft Fabric** (Lakehouse/Warehouse patterns) to support governed reporting and self-service analytics.
* Build and manage curated data layers aligned to medallion-style processing (raw → standardized → curated) using
** Spark/PySpark**,
** SQL**, and
** Python**.
* Develop and maintain enterprise data models optimized for analytics performance, consistent KPI definitions, and reuse across business domains.
** Data Integration & Ingestion (Infor OS Data Lake + APIs + Flows)
*** Develop and support automated ingestion from
** Infor OS Data Lake
** using supported extraction/integration methods such as: +
** Data Fabric Objects API** (object/file extraction) +
** Compass SQL API / Compass JDBC Driver** (query-based extraction) +
** ION Data Lake Flows** (scheduled push to connection points) +
** ETL Client for Data Lake** (scheduled transfer patterns) +
** Stream Pipelines
** where applicable for continuous/near real-time delivery
* Implement incremental loading patterns, orchestration, monitoring, alerting, and failure recovery to ensure reliable delivery of daily/near real-time datasets.
* Partner with application and integration teams to align ingestion with upstream interfaces, data contracts, and security requirements.
** Reporting & Analytics Enablement (Power BI + Dataset Modeling)
*** Provide trusted, well-documented datasets that enable enterprise dashboards and self-service analytics in
** Power BI**.
* Build and maintain business-friendly semantic/dimensional models that support high-performance dashboards and consistent KPI definitions.
* Support modernization and migration of reporting assets into
** Microsoft Fabric**, ensuring datasets and models align to reporting needs and enterprise metric definitions.
** Data Quality, Governance & Master Data Support
*** Implement validation, reconciliation, and anomaly detection to ensure accuracy and completeness of curated datasets.
* Establish automated checks for common data issues (duplicates, missing attributes, invalid statuses, inconsistent units of measure).
* Partner with master data stakeholders and business data stewards to define standards, drive adoption, and remediate root-cause issues impacting data quality.
** AI-Enabled Analytics Support (Practical AI Competency)
*** Prepare AI-ready datasets by ensuring data completeness, consistency, lineage, and documentation (e.g., feature-ready curated tables, standardized definitions, and auditability).
* Support AI-assisted development workflows (e.g., using copilots/assistants to accelerate transformation code generation, documentation, and standardization) while adhering to company AI requirements for accuracy, confidentiality, compliance, and labeling of AI-generated content where required.
** Operational Excellence (Reliability, Supportability, Documentation)
*** Develop and…
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
Search for further Jobs Here:
×