Roles & Responsibilities
Key Responsibilities :
- Design, develop, and maintain data pipelines using PySpark on Databricks.
- Write clean, efficient, and reusable PySpark code for data ingestion, transformation, and processing.
- Conduct unit and integration testing of data workflows to ensure accuracy and performance.
- Collaborate with data scientists, analysts, and other engineers to support data-driven initiatives.
- Optimize data workflows for scalability and performance in cloud environments.
- Document technical solutions and maintain code repositories.
Qualifications :
Bachelor’s degree in Computer Science, Engineering, or related field.3+ years of experience in data engineering, with excellent proficiency in PySpark.Hands-on experience with Databricks platform.Familiarity with data testing frameworks and best practices.Good understanding of ETL processes, data modeling, and performance tuning.Interested Applicants, please email your resume to rupa.kankariya@manpower.com.sg (R1441955), stating the position as the subject title in the email. All Applications will be handled with strict confidentiality.
Tell employers what skills you have
Microsoft Azure
PySpark
Scalability
Azure
Data Modeling
Pipelines
ETL
Data Engineering
Performance Tuning
Cloud
Integration Testing