Databricks SME and AWS Data Engineer

Databricks SME and AWS Data Engineer

Posted Today by 1771062725

£480 Per day
Inside
Hybrid
Northampton

Summary: The role of Databricks SME and AWS Data Engineer involves migrating data from Aerospike to Postgres and utilizing Databricks for back-testing fraud detection models. The position requires collaboration with cross-functional teams to develop and deliver Proof of Concepts (PoCs) while ensuring seamless integration with existing systems. The candidate will lead ETL and data engineering work on AWS and take ownership of tasks from the backlog. This is a hybrid position based in Northampton, UK, for a duration of 6 months.

Key Responsibilities:

  • Collaborate with cross-functional teams to architect, design, develop and deliver PoCs related to fraud detection.
  • Lead the ETL & Data manipulation/engineering work on AWS.
  • Integrate Databricks-based back-testing into the fraud detection pipeline.
  • Work closely with architects and other developers to ensure seamless integration with existing systems.
  • Participate in weekly stand-up calls to demonstrate progress and align on deliverables.
  • Take ownership of tasks from the backlog, ensuring timely and high-quality delivery.

Key Skills:

  • Strong AWS Data Engineering expertise.
  • Proficiency in Kafka for real-time data streaming and integration.
  • Proficiency with Databricks for data processing and analytics.
  • Working knowledge of NodeJS would be an added advantage.
  • Solid programming skills in Python, PySpark, Spark.
  • Candidate must be adept at working with large-scale datasets, S3, Python, and data cataloguing tools. Familiarity with data engineering best practices is essential.
  • Performance and Code Quality - candidate should demonstrate a strong commitment to building high-performance, scalable, and resilient distributed systems, with an emphasis on clean, maintainable, and testable code.
  • CI/CD Proficiency - Hands-on experience with CI/CD pipelines and tools (e.g., Jenkins, GitHub Actions, GitLab CI, etc.) for automated build, test, and deployment processes.
  • Secure Development Practices - Awareness of secure coding standards, data protection principles, and experience working in regulated environments (especially relevant for fraud detection and financial services).
  • Testing Rigor - Strong understanding of unit testing, integration testing, and test automation frameworks (e.g., PyTest, etc.) to ensure code quality and reliability.
  • Familiarity with cloud-native development and CI/CD practices.
  • Agile mindset with a proactive, developer-driven approach to problem-solving.

Salary: £480.00 per day

City: Northampton

Country: UK

Working Arrangements: hybrid

IR35 Status: inside IR35

Seniority Level: undetermined

Industry: IT

Detailed Description From Employer:

Databricks SME and AWS Data Engineer
Location: UK - Northampton (Hybrid)
6 months

UMBRELLA ONLY

Project Objective:
A key initiative involves migrating from Aerospike to Postgres and leveraging Databricks for back-testing fraud detection models. This role will contribute to the development and integration of Proof of Concepts (PoCs) from the detection backlog.

Key Responsibilities:

  • Collaborate with cross-functional teams to architect, design, develop and deliver PoCs related to fraud detection.
  • Lead the ETL & Data manipulation/engineering work on AWS.
  • Integrate Databricks-based back-testing into the fraud detection pipeline.
  • Work closely with architects and other developers to ensure seamless integration with existing systems.
  • Participate in weekly stand-up calls to demonstrate progress and align on deliverables.
  • Take ownership of tasks from the backlog, ensuring timely and high-quality delivery.


Required Skills & Experience:

  • Strong AWS Data Engineering expertise.
  • Proficiency in Kafka for real-time data streaming and integration.
  • Proficiency with Databricks for data processing and analytics.
  • Working knowledge of NodeJS would be an added advantage.
  • Solid programming skills in Python, PySpark, Spark.
  • Candidate must be adept at working with large-scale datasets, S3, Python, and data cataloguing tools. Familiarity with data engineering best practices is essential.
  • Performance and Code Quality - candidate should demonstrate a strong commitment to building high-performance, scalable, and resilient distributed systems, with an emphasis on clean, maintainable, and testable code.
  • CI/CD Proficiency - Hands-on experience with CI/CD pipelines and tools (e.g., Jenkins, GitHub Actions, GitLab CI, etc.) for automated build, test, and deployment processes.
  • Secure Development Practices - Awareness of secure coding standards, data protection principles, and experience working in regulated environments (especially relevant for fraud detection and financial services).
  • Testing Rigor - Strong understanding of unit testing, integration testing, and test automation frameworks (e.g., PyTest, etc.) to ensure code quality and reliability.
  • Familiarity with cloud-native development and CI/CD practices.
  • Agile mindset with a proactive, developer-driven approach to problem-solving.


Ideal Candidate Profile:

  • A hands-on architect & developer with a strong sense of ownership and accountability.
  • Comfortable working in a collaborative, fast-paced environment.
  • Able to pick up tasks independently, contribute to design discussions, and deliver integrated solutions.
  • Strong communication skills to engage with both technical and non-technical stakeholders.

All profiles will be reviewed against the required skills and experience. Due to the high number of applications we will only be able to respond to successful applicants in the first instance. We thank you for your interest and the time taken to apply!