More jobs:
PySpark Developer
Job in
Hartford, Hartford County, Connecticut, 06112, USA
Listed on 2025-11-19
Listing for:
Inizio Partners Corp
Full Time
position Listed on 2025-11-19
Job specializations:
-
IT/Tech
Data Engineer, Big Data
Job Description & How to Apply Below
We are seeking a highly skilled and experienced Python and PySpark Developer to join our team. The ideal candidate will be responsible for designing, developing, and optimizing big data pipelines and solutions using Python, PySpark, and distributed computing frameworks. This role involves working closely with data engineers, data scientists, and business stakeholders to process, analyze, and derive insights from large-scale datasets.
Key Responsibilities- Design and implement scalable data pipelines using PySpark and other big data frameworks.
- Develop reusable and efficient code for data extraction, transformation, and loading (ETL).
- Optimize data workflows for performance and cost efficiency.
- Process and analyze structured and unstructured datasets.
- Build and maintain data lakes, data warehouses, and other storage solutions.
- Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions.
- Troubleshoot and resolve performance bottlenecks in big data pipelines.
- Write clean, maintainable, and well-documented code.
- Ensure compliance with data governance and security policies.
Skills & Qualifications Programming Skills
- Proficient in Python with experience in data processing libraries like Pandas and Num Py.
- Strong experience with PySpark and Apache Spark.
- Hands‑on experience with big data platforms such as Hadoop, Databricks, or similar.
- Familiarity with cloud services like AWS (EMR, S3), Azure (Data Lake, Synapse), or Google Cloud (Big Query, Dataflow).
- Strong knowledge of SQL and No
SQL databases. - Experience working with relational databases like Postgre
SQL, MySQL, or Oracle.
- Experience with workflow orchestration tools like Apache Airflow or similar.
- Ability to solve complex data engineering problems efficiently.
- Strong communication skills to work effectively in a collaborative environment.
- Knowledge of data Lakehouse architectures and frameworks.
- Familiarity with machine learning pipelines and integration.
- Experience in CI/CD tools and Dev Ops practices for data workflows.
- Certification in Spark, Python, or cloud platforms is a plus.
- Bachelors or Masters degree in Computer Science, Data Engineering, or a related field.
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
Search for further Jobs Here:
×