£800 Per day
Inside
Hybrid
City of London, London
Summary: The PySpark Engineer - Data Specialist role involves developing and optimizing data processing pipelines using PySpark within a Tier 1 investment banking environment. The position requires collaboration with cross-functional teams to automate ETL processes and ensure data integrity and security. The contract is hybrid and spans six months, with an immediate start date. The role is classified as inside IR35, requiring engagement through an umbrella company.
Key Responsibilities:
- Develop, maintain, and optimize PySpark data processing pipelines in a fast-paced investment banking environment.
- Automate ETL processes (data extraction, transformation, and loading) to ensure seamless data flow across systems.
- Collaborate with cross-functional teams, including data engineers and analysts, to implement data-driven solutions tailored for investment banking needs.
- Leverage PySpark and Apache Spark to efficiently handle large datasets and improve processing efficiency.
- Optimize SQL queries for faster data retrieval and integration across banking systems.
- Ensure data integrity, quality, and security throughout the data pipeline lifecycle.
- Troubleshoot and resolve data-related issues to maintain seamless reporting and analytics workflows.
Key Skills:
- Proven experience in data processing and automation within an investment banking environment.
- Strong proficiency in PySpark and Apache Spark for data pipeline development.
- Solid understanding of SQL and experience optimizing complex queries.
- Expertise in automating ETL processes to improve data flow and efficiency.
- Excellent problem-solving skills, attention to detail, and ability to manage complex datasets.
- Strong communication skills with the ability to work in a collaborative, fast-paced team environment.
Salary (Rate): 800
City: London
Country: United Kingdom
Working Arrangements: hybrid
IR35 Status: inside IR35
Seniority Level: undetermined
Industry: IT