£510 Per day
Inside
Remote
City of London
Summary: The Data Analyst role focuses on understanding and documenting existing systems to support decommissioning activities. The position involves analyzing solutions built with Java, Node JS, and React, and creating a comprehensive overview of data flows within the program. Key responsibilities include documenting data flows, system dependencies, and ensuring data quality management. The role requires strong analytical skills and experience in cloud environments, particularly AWS or Azure.
Key Responsibilities:
- Analyze current solutions built using Java, Node JS, and React.
- Document data flows, system dependencies, and underlying data models.
- Investigate daily system usage and clarify ownership and integration points.
- Utilize Python and PySpark to analyze data pipelines.
- Ensure accuracy and trust in documented data flows and models through testing and data quality management.
- Work in cloud environments such as AWS or Azure.
Key Skills:
- Java background
- Node JS
- Json
- RDS
- React
- Data Modelling
- Python / Spark
- Cloud experience (AWS / Azure)
- AWS Glue
- Databricks
- Testing e.g. PyTest
- Data Quality e.g. Great Expectations
Salary (Rate): £510 per day
City: City of London
Country: United Kingdom
Working Arrangements: remote
IR35 Status: inside IR35
Seniority Level: undetermined
Industry: IT
Data Analyst
Onsite Requirements: Remote
Start Date: ASAP
Role Duration: 1 year
Clerance Requirements: Active SC clearance
Inside IR35 - umbrella only
Role Description:
We're looking for a Data Engineer whose main focus is understanding and documenting existing systems, with the goal of supporting decommissioning activities. The role centres on analysing current solutions built using Java, Node JS, and React, and developing a clear, end to end picture of how data flows across the wider programme.
This includes documenting data flows, system dependencies, and underlying data models, ensuring there is a clear record of how data is structured, stored, and used throughout the solution. The role involves investigating how systems are used on a day-to-day basis, clarifying ownership and integration points, and capturing this information in a way that supports risk assessment and decommissioning decisions.
Responsibilities:
Python and PySpark are required as supporting capabilities, used where needed to analyse data pipelines and confirm how data moves and transforms in practice. The role also requires strong experience with testing and data quality management, ensuring that documented data flows and models are accurate and trusted. Experience working in cloud environments such as AWS or Azure is expected, with Databricks considered a nice to have.
Required Skills:
- Java background
- Node JS
- Json
- RDS
- React
- Data Modelling
- Python / Spark
- Cloud experience (AWS / Azure) o AWS Glue o Databricks
- Testing e.g. PyTest
- Data Quality e.g. Great Expectations