Negotiable
Undetermined
Remote
Remote or San Jose, California
Summary: The role of Senior Python Developer / Data Engineer focuses on designing and building large-scale data pipelines, managing ETL/ELT workflows, and developing end-to-end machine learning pipelines. The position requires collaboration with data scientists to productionize models and ensure data quality and pipeline reliability. This is a fully remote position, allowing for flexibility in work arrangements. Candidates should have extensive experience in software engineering, particularly with Python and data engineering practices.
Key Responsibilities:
- Design and build large-scale data pipelines for ingestion, transformation and processing
- Work on ETL/ELT workflows handling different types of data
- Build and maintain end-to-end ML pipelines from data preparation to deployment and monitoring
- Collaborate with data scientists to productionize ML models
- Work on feature engineering, training pipelines and model serving
- Ensure data quality, monitoring and pipeline reliability
- Optimize systems for performance, scalability and cost
- Contribute to clean, maintainable, production-grade Python code
Key Skills:
- 8+ years of software engineering experience with Python as primary language
- Strong background in data engineering (ETL/ELT, pipelines, data processing)
- Hands-on experience building and maintaining ML pipelines in production environments
- Experience with PySpark / Apache Spark
- Experience with workflow orchestration tools like Airflow, Dagster, or Prefect
- Good understanding of streaming/data processing systems (Kafka, Kinesis, etc.)
- Experience working with cloud platforms (AWS / Google Cloud Platform / Azure)
- Strong SQL skills and experience with data warehouses
- Comfortable working in a distributed/remote engineering setup
Salary (Rate): £51,000 yearly
City: undetermined
Country: undetermined
Working Arrangements: remote
IR35 Status: undetermined
Seniority Level: undetermined
Industry: IT
Senior Python Developer / Data Engineer / ML Pipelines
100% Remote role
Responsibilities
-
Design and build large-scale data pipelines for ingestion, transformation and processing
-
Work on ETL/ELT workflows handling different types of data
-
Build and maintain end-to-end ML pipelines from data preparation to deployment and monitoring
-
Collaborate with data scientists to productionize ML models
-
Work on feature engineering, training pipelines and model serving
-
Ensure data quality, monitoring and pipeline reliability
-
Optimize systems for performance, scalability and cost
-
Contribute to clean, maintainable, production-grade Python code
Experience Required
-
8+ years of software engineering experience with Python as primary language
-
Strong background in data engineering (ETL/ELT, pipelines, data processing)
-
Hands-on experience building and maintaining ML pipelines in production environments
-
Experience with PySpark / Apache Spark
-
Experience with workflow orchestration tools like Airflow, Dagster, or Prefect
-
Good understanding of streaming/data processing systems (Kafka, Kinesis, etc.)
-
Experience working with cloud platforms (AWS / Google Cloud Platform / Azure)
-
Strong SQL skills and experience with data warehouses
-
Comfortable working in a distributed/remote engineering setup
Plus
-
Experience with NLP or LLM-based systems
-
Familiarity with MLOps tools like MLflow, Kubeflow, or similar
-
Experience with feature stores
-
Exposure to data privacy, PII detection, or compliance-related systems