Data Engineer AWS / PySpark / Glue

Data Engineer AWS / PySpark / Glue

Posted 3 days ago by Teqdata

Negotiable
Undetermined
Remote
Remote

Summary: We are seeking an experienced Data Engineer with expertise in AWS cloud technologies, Python, SQL, and big data processing. The role involves designing, developing, and optimizing ETL workflows, as well as building scalable data pipelines and collaborating with cross-functional teams. Hands-on experience with AWS Glue and Terraform is essential for managing infrastructure and supporting large-scale data processing. The ideal candidate will deliver reliable data solutions in a cloud-native environment.

Key Responsibilities:

  • Design and develop scalable ETL/data pipelines
  • Build and optimize data processing workflows using PySpark
  • Work extensively with AWS services including AWS Glue
  • Develop and maintain infrastructure using Terraform
  • Write efficient SQL queries and Python scripts for data transformation and automation
  • Monitor, troubleshoot, and improve data platform performance
  • Collaborate with engineering and analytics teams to support business requirements

Key Skills:

  • Expertise in AWS cloud technologies
  • Proficiency in Python and SQL
  • Experience with big data processing and ETL workflows
  • Hands-on experience with AWS Glue and PySpark
  • Knowledge of infrastructure management using Terraform
  • Strong problem-solving and troubleshooting skills
  • Ability to collaborate with cross-functional teams

Salary (Rate): £45

City: undetermined

Country: undetermined

Working Arrangements: remote

IR35 Status: undetermined

Seniority Level: undetermined

Industry: IT

Detailed Description From Employer:

We are seeking an experienced Data Engineer with strong expertise in AWS cloud technologies, Python, SQL, and big data processing. The ideal candidate will have hands-on experience building scalable data pipelines, working with AWS Glue, and managing infrastructure using Terraform.

In this role, you will design, develop, and optimize ETL workflows, support large-scale data processing using PySpark, and collaborate with cross-functional teams to deliver reliable data solutions in a cloud-native environment

Responsibilities:

Design and develop scalable ETL/data pipelines

Build and optimize data processing workflows using PySpark

Work extensively with AWS services including AWS Glue

Develop and maintain infrastructure using Terraform

Write efficient SQL queries and Python scripts for data transformation and automation

Monitor, troubleshoot, and improve data platform performance

Collaborate with engineering and analytics teams to support business requirements