£510 Per day
Inside
Remote
City of London
Summary: The Data Engineer role focuses on understanding and documenting existing systems to support decommissioning activities. The position involves analyzing solutions built with Java, Node JS, and React, and creating a comprehensive overview of data flows within the program. Key responsibilities include documenting data flows, system dependencies, and ensuring data quality management. The role requires active SC clearance and is classified as inside IR35, with an emphasis on cloud experience and data analysis capabilities.
Key Responsibilities:
- Understand and document existing systems to support decommissioning activities.
- Analyze current solutions built using Java, Node JS, and React.
- Document data flows, system dependencies, and underlying data models.
- Investigate daily system usage, clarify ownership and integration points.
- Utilize Python and PySpark to analyze data pipelines.
- Ensure accuracy and trust in documented data flows and models through testing and data quality management.
- Work in cloud environments such as AWS or Azure.
Key Skills:
- Java background
- Node JS
- Json
- RDS
- React
- Data Modelling
- Python / Spark
- Cloud experience (AWS / Azure)
- AWS Glue
- Databricks
- Testing e.g. PyTest
- Data Quality e.g. Great Expectations
Salary (Rate): £510 per day
City: City of London
Country: United Kingdom
Working Arrangements: remote
IR35 Status: inside IR35
Seniority Level: undetermined
Industry: IT
Data Engineer
Onsite Requirements: Remote
Start Date: ASAP
Role Duration: 1 year
Clerance Requirements: Active SC clearance
Inside IR35 - umbrella only
Role Description:
We're looking for a Data Engineer whose main focus is understanding and documenting existing systems, with the goal of supporting decommissioning activities. The role centres on analysing current solutions built using Java, Node JS, and React, and developing a clear, end to end picture of how data flows across the wider programme.
This includes documenting data flows, system dependencies, and underlying data models, ensuring there is a clear record of how data is structured, stored, and used throughout the solution. The role involves investigating how systems are used on a day-to-day basis, clarifying ownership and integration points, and capturing this information in a way that supports risk assessment and decommissioning decisions.
Responsibilities:
Python and PySpark are required as supporting capabilities, used where needed to analyse data pipelines and confirm how data moves and transforms in practice. The role also requires strong experience with testing and data quality management, ensuring that documented data flows and models are accurate and trusted. Experience working in cloud environments such as AWS or Azure is expected, with Databricks considered a nice to have.
Required Skills:
- Java background
- Node JS
- Json
- RDS
- React
- Data Modelling
- Python / Spark
- Cloud experience (AWS / Azure) o AWS Glue o Databricks
- Testing e.g. PyTest
- Data Quality e.g. Great Expectations