Machine Learning Engineer

Machine Learning Engineer

Posted 1 day ago by Russell Tobin

£550 Per day
Outside
Hybrid
London Area, United Kingdom

Summary: The role of Machine Learning Engineer focuses on designing and deploying AI agents and workflows powered by large language models (LLMs) in a production-first environment. Candidates will be responsible for end-to-end project ownership, integrating AI systems into operational workflows, and ensuring scalability and reliability. The position emphasizes practical implementation and measurable business impact rather than experimental work. Ideal candidates will have a strong technical background and experience in deploying AI systems at scale.

Key Responsibilities:

  • Designing and building AI agents and agentic workflows powered by LLMs
  • Developing systems using RAG, reasoning, planning, memory, and tool orchestration
  • Building multi-step intelligent systems capable of real-world tool usage
  • Working with MCP-style architectures to structure context and improve interoperability
  • Contributing to recommendation, classification, and forecasting systems using large-scale datasets
  • Automating business workflows and decision-making processes through AI-driven solutions
  • Owning projects end-to-end from concept through to production deployment and iteration
  • Building and deploying AI agents that operate reliably in production environments
  • Integrating AI systems into APIs, products, and operational workflows
  • Collaborating closely with engineering teams to ensure scalability, observability, and maintainability
  • Making pragmatic decisions balancing model performance, latency, and cost efficiency

Key Skills:

  • Strong Python skills with experience writing production-grade code
  • Proven experience deploying LLM-powered systems into production environments
  • Hands-on experience with LangChain, LangGraph, or equivalent orchestration frameworks
  • Experience building AI agents and agentic workflows with tool usage and multi-step reasoning
  • Strong understanding and implementation experience of RAG systems
  • Familiarity with MCP/FastMCP/FastAPI or similar orchestration patterns
  • Strong understanding of LLM trade-offs including hallucination mitigation, latency, and cost optimisation
  • Experience deploying AI systems in cloud environments such as AWS, GCP, or Azure
  • Working knowledge of SQL/data manipulation
  • Experience working on SaaS or B2B AI products or delivering AI-driven transformation within an organisation
  • A background in high-growth or scaling environments
  • Clear evidence of systems that are actively used and delivering value

Salary (Rate): £550 daily

City: London

Country: United Kingdom

Working Arrangements: hybrid

IR35 Status: outside IR35

Seniority Level: undetermined

Industry: IT

Detailed Description From Employer:

Data Scientist / Machine Learning Scientist

Location: London (Hybrid)

Contract: Outside IR35

Rate: £500–£550 per day (depending on interview outcome)

We’re looking for AI operators who ship — not experiment. This is an opportunity to join a major AI build focused on deploying real-world LLM and agentic systems at scale across both AI products and enterprise transformation initiatives. You’ll be working in a production-first environment where the emphasis is on building reliable, scalable AI systems that deliver measurable business impact.

What You’ll Be Working On

  • Designing and building AI agents and agentic workflows powered by LLMs
  • Developing systems using RAG, reasoning, planning, memory, and tool orchestration
  • Building multi-step intelligent systems capable of real-world tool usage
  • Working with MCP-style architectures (or equivalent) to structure context and improve interoperability
  • Contributing to recommendation, classification, and forecasting systems using large-scale datasets
  • Automating business workflows and decision-making processes through AI-driven solutions

What You’ll Be Doing

  • Owning projects end-to-end from concept through to production deployment and iteration
  • Building and deploying AI agents that operate reliably in production environments
  • Integrating AI systems into APIs, products, and operational workflows
  • Collaborating closely with engineering teams to ensure scalability, observability, and maintainability
  • Making pragmatic decisions balancing model performance, latency, and cost efficiency

Core Requirements

  • Strong Python skills with experience writing production-grade code
  • Proven experience deploying LLM-powered systems into production environments
  • Hands-on experience with LangChain, LangGraph, or equivalent orchestration frameworks
  • Experience building AI agents and agentic workflows with tool usage and multi-step reasoning
  • Strong understanding and implementation experience of RAG systems
  • Familiarity with MCP/FastMCP/FastAPI or similar orchestration patterns
  • Strong understanding of LLM trade-offs including hallucination mitigation, latency, and cost optimisation
  • Experience deploying AI systems in cloud environments such as AWS, GCP, or Azure
  • Working knowledge of SQL/data manipulation ( Working knowledge of SQL or data manipulation is expected, but it is not a primary focus for this role.)

Strong signals include:

  • Experience working on SaaS or B2B AI products or delivering AI-driven transformation within an organisation.
  • A background in high-growth or scaling environments , where speed and pragmatism are critical.
  • Clear evidence of systems that are actively used and delivering value , rather than experimental work.

Ideal Background

  • Masters degree or higher in Computer Science, Mathematics, Engineering, or a related technical field
  • Experience building and iterating on AI systems delivering measurable value
  • Strong ownership mindset and ability to operate in fast-moving environments
  • Product-focused approach with a bias toward delivering impact

Why This Role

  • Work on live AI systems used at scale
  • Join a well-supported AI engineering environment
  • High ownership and visibility across products and operations
  • Opportunity to shape enterprise AI adoption in a meaningful way