More jobs:
GCP Data Engineer
Job in
Frisco, Collin County, Texas, 75034, USA
Listed on 2026-01-01
Listing for:
Sunixa Solutions Inc.
Full Time
position Listed on 2026-01-01
Job specializations:
-
IT/Tech
Data Engineer -
Engineering
Data Engineer
Job Description & How to Apply Below
Expires on January 29, 2026
Role: GCP Data Engineer
Location:
Frisco, TX (Hybrid)
Duration:
Long term
We are seeking a skilled Google Cloud Platform Data Engineer to design, build, and optimise data pipelines and analytics solutions in the cloud. The ideal candidate must have hands‑on experience with Google Cloud Platform data services, strong ETL/ELT development skills, and a solid understanding of data architecture, data modelling, data warehousing and performance optimisation.
Key Responsibilities- Develop ETL/ELT processes to extract data from various sources, transform it, and load it into Big Query or other target systems.
- Build and maintain data models, data warehouses, and data lakes for analytics and reporting.
- Design and implement scalable, secure, and efficient data pipelines on Google Cloud Platform using tools such as Dataflow, Pub/Sub, Cloud Run, Python and Linux scripting.
- Optimise Big Query queries, manage partitioning and clustering, and handle cost optimisation.
- Integrate data from on-premise and cloud systems using Cloud Storage, and APIs.
- Work closely with Dev Ops teams to automate deployments using Terraform, Cloud Build, or CI/CD pipelines.
- Ensure security and compliance by applying IAM roles, encryption, and network controls.
- Collaborate with data analysts, data scientists, and application teams to deliver high-quality data solutions.
- Implement best practices for data quality, monitoring, and governance.
- Bachelor’s degree in Computer Science, Information Technology, or related field.
- Experience in data engineering, preferably in a cloud environment.
- Hands‑on and strong expertise in Google Cloud Platform services:
Big Query, Cloud Storage, Cloud Run, Dataflow, Cloud SQL, Alloy
DB, Cloud Balancer, Pub/Sub, IAM, Logging and Monitoring. - Proficiency in SQL, Python and Linux scripting.
- Prior experience with ETL tools such as Datastage, Informatica, SSIS.
- Familiarity with data modeling (star/snowflake) and data warehouse concepts.
- Understanding of CI/CD, version control (Git), and Infrastructure as Code (Terraform).
- Strong problem‑solving and analytical mindset.
- Effective communication and collaboration skills.
- Ability to work in an agile and fast‑paced environment.
- Google Cloud Platform Professional Data Engineer or Cloud Architect certification is a plus.
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
Search for further Jobs Here:
×