Negotiable
Undetermined
Remote
Remote
Summary: We are seeking an experienced Data Engineer with expertise in AWS cloud technologies, Python, SQL, and big data processing. The role involves designing, developing, and optimizing ETL workflows, as well as building scalable data pipelines and collaborating with cross-functional teams. Hands-on experience with AWS Glue and Terraform is essential for managing infrastructure and supporting large-scale data processing. The ideal candidate will deliver reliable data solutions in a cloud-native environment.
Key Responsibilities:
- Design and develop scalable ETL/data pipelines
- Build and optimize data processing workflows using PySpark
- Work extensively with AWS services including AWS Glue
- Develop and maintain infrastructure using Terraform
- Write efficient SQL queries and Python scripts for data transformation and automation
- Monitor, troubleshoot, and improve data platform performance
- Collaborate with engineering and analytics teams to support business requirements
Key Skills:
- Expertise in AWS cloud technologies
- Proficiency in Python and SQL
- Experience with big data processing and ETL workflows
- Hands-on experience with AWS Glue and PySpark
- Knowledge of infrastructure management using Terraform
- Strong problem-solving and troubleshooting skills
- Ability to collaborate with cross-functional teams
Salary (Rate): £45
City: undetermined
Country: undetermined
Working Arrangements: remote
IR35 Status: undetermined
Seniority Level: undetermined
Industry: IT
We are seeking an experienced Data Engineer with strong expertise in AWS cloud technologies, Python, SQL, and big data processing. The ideal candidate will have hands-on experience building scalable data pipelines, working with AWS Glue, and managing infrastructure using Terraform.
In this role, you will design, develop, and optimize ETL workflows, support large-scale data processing using PySpark, and collaborate with cross-functional teams to deliver reliable data solutions in a cloud-native environment
Responsibilities:
Design and develop scalable ETL/data pipelines
Build and optimize data processing workflows using PySpark
Work extensively with AWS services including AWS Glue
Develop and maintain infrastructure using Terraform
Write efficient SQL queries and Python scripts for data transformation and automation
Monitor, troubleshoot, and improve data platform performance
Collaborate with engineering and analytics teams to support business requirements