More jobs:
Sr. Data Platform Engineer
Job in
Elk Grove, Sacramento County, California, 95759, USA
Listed on 2025-12-09
Listing for:
ALLDATA
Full Time
position Listed on 2025-12-09
Job specializations:
-
IT/Tech
Data Engineer, Cloud Computing
Job Description & How to Apply Below
Join to apply for the Sr. Data Platform Engineer role at ALLDATA
Job DescriptionHybrid role 3X a week in office in Elk Grove, CA; no remote capabilities
The Senior Data Platform Engineer is a technical expert responsible for designing, implementing, and supporting scalable data solutions using modern big data technologies. In this full-time role, you will work closely with data architects, analysts, and business stakeholders to build, optimize, and maintain data pipelines and platforms that drive analytics, reporting, and machine learning. You will ensure the reliability, performance, and security of production data systems, and play a key role in troubleshooting, monitoring, and continuous improvement.
PositionResponsibilities – Other Duties May Be Assigned
- Design, develop, and maintain robust data pipelines for ingesting, transforming, and storing large volumes of structured, semi-structured, and unstructured data.
- Implement data workflows and ETL processes using technologies such as Spark, Delta Lake, and cloud-native tools.
- Support the production use of big data platforms (e.g., Databricks, Snowflake, Google Cloud Platform), ensuring high availability, scalability, and performance.
- Monitor, troubleshoot, and resolve issues in production data systems, including job failures, performance bottlenecks, and data quality concerns.
- Collaborate with data architects to translate business requirements into technical solutions, ensuring alignment with architectural standards and best practices.
- Optimize SQL queries, Spark jobs, and data processing workloads for efficiency and cost-effectiveness.
- Implement and maintain data governance, security, and compliance measures, including access controls, data masking, and audit logging.
- Integrate data workflows into CI/CD pipelines, automate deployment processes, and manage source control using Git and related Dev Ops tools.
- Document data pipelines, workflows, and operational procedures to support knowledge sharing and maintainability.
- Stay current with emerging big data technologies and recommend improvements to enhance reliability, scalability, and efficiency.
- 7+ years of experience in data engineering or related roles, with hands‑on experience building and supporting production data pipelines.
- Strong proficiency with big data platforms (e.g., Databricks, Snowflake, Google Cloud Platform) and the Spark ecosystem.
- Experience with data lakehouse and warehouse architectures, including Delta Lake or similar technologies.
- Senior Data Engineer with deep expertise in Retrieval‑Augmented Generation (RAG) systems and advanced chunking strategies for optimizing semantic search and LLM integration.
- Advanced skills in PySpark, Python, SQL, and related data processing frameworks.
- Experience with cloud storage, messaging/streaming technologies (e.g., Apache Kafka, cloud Pub/Sub), and vector databases.
- Proven ability to optimize ETL jobs, tune cluster configurations, and troubleshoot production issues.
- Familiarity with data governance tools, catalog solutions, and security best practices.
- Experience integrating data workflows into CI/CD pipelines and using Dev Ops tools (e.g., Git, Jenkins, Terraform).
- Strong problem‑solving skills, attention to detail, and ability to work independently or as part of a team.
- Excellent communication skills to collaborate with cross‑functional teams and document technical solutions.
- Experience working in an Agile environment.
- Big Data Platforms:
Spark, Delta Lake, SQL engines, MLflow (or similar model tracking tools), relational databases. - Data Governance:
Data catalog and access control solutions. - Programming & Data Processing:
PySpark, Python, SQL. - Cloud Services:
Cloud storage, messaging/streaming technologies, vector databases, Big Query. - Dev Ops & CI/CD:
Git for version control, Jenkins, infrastructure‑as‑code tools (e.g., Terraform). - Other Tools:
Project and workflow management tools (e.g., JIRA, Confluence), BI/analytics tools (e.g., Power BI, Looker Studio).
- Certifications in major big data platforms (e.g., Databricks, Snowflake, Google Cloud Platform) are advantageous.
- Exposure to related big data and…
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
Search for further Jobs Here:
×