MLOps Engineer: ML Risk Platform

WhyHireWrong?
London
4 days ago
Create job alert

Every day, the models make financial decisions that affect real people. Credit approvals, fraud blocks, transaction risk scores. If a model drifts silently in production, customers get wrongly declined. If a pipeline breaks at 2am, no one catches it until the damage is done.

This role exists to make sure that does not happen.

You will own the infrastructure that takes ML models from a data scientist's notebook into production systems processing millions of events daily, and keeps them running reliably across multiple regulatory jurisdictions. Not maintaining someone else's setup. Building and owning it.

What You Will Work On

  • Model pipelines: Design and operate automated training, validation, deployment, and rollback workflows across our credit scoring, fraud detection, and transaction risk models

  • Production monitoring: Build observability for ML specific failure modes including data drift, prediction drift, and feature skew, not just system uptime

  • Compliance instrumentation: Maintain full audit trails and model cards required for internal model risk reviews and regulatory examination under EU AI Act and GDPR

  • Infrastructure ownership: Run Kubernetes based ML serving on AWS or Azure, manage CI/CD pipelines that version code, data, and models simultaneously

  • Reliability and incident response: Define SLAs for latency sensitive scoring models and own the full response when something breaks in production

  • Cost management: Optimise cloud spend for GPU training jobs and batch inference workloads, compute budgets in fintech are scrutinised closely

    5 Non-Negotiable Requirements

1. Production ML pipelines you built yourself
You have designed and operated automated training, validation, and deployment pipelines serving real users in a live environment. Not internal tooling. Not a prototype. If the pipeline broke, you were the one who fixed it.

2. Kubernetes in production
You have deployed and managed containerised ML workloads on Kubernetes including autoscaling, resource limits, and failure recovery. EKS, AKS, or GKE.

3. ML lifecycle ownership
Hands on model versioning, experiment tracking, and registry management using MLflow, Weights and Biases, or equivalent. You managed promotion gates and rollback procedures, not just tracked experiments.

4. Monitoring for ML specific failures
You have built observability for data drift, prediction drift, and feature skew, not just CPU and memory. Evidently AI, Whylogs, Prometheus, or equivalent. You defined what an alert means and what to do when it fires.

5. Regulated environment experience
You have worked in fintech, banking, or insurance where model decisions required audit trails, explainability artefacts, or sign off from a risk or compliance function. You know what SR 11-7, EU AI Act, or GDPR means for an ML pipeline in practice.

Full Technical Stack

Core: Python, Docker, Kubernetes, GitHub Actions or GitLab CI

ML Platform: MLflow, Apache Airflow or Prefect

Cloud: AWS SageMaker with EKS, or Azure ML with AKS

Monitoring: Prometheus, Grafana, Evidently AI

Data: Spark, PostgreSQL, S3 or Azure Blob

Useful but not required on day one: Terraform, feature stores such as Feast or Tecton, LangChain for LLM pipeline integration, SHAP or LIME for explainability

What This Role Is Not

Not a data science role. You will not be building models.

Not a generic DevOps role. Kubernetes experience without ML context is not sufficient.

Not a research or platform architecture role. All work is production focused with hard reliability and compliance constraints.

How to Apply

This role is open to EU based candidates only. We are not considering applications from outside the European Union at this time, regardless of remote working arrangements or timezone compatibility.

Submit your CV and record a short video answer to one question:

Describe a machine learning pipeline you built and owned in production. What broke, how did you detect it, and what did you change?

The video format is uncomfortable. We know that. If you still do it, that already tells us something.

How Applications Are Assessed

I want to be upfront about how this works before you invest your time.

Every CV is scored against the 5 non-negotiable requirements only. One point per requirement. 5 out of 5 to proceed. Not 4. If a requirement is listed as a tool or skill without context describing what you built and what it served, it scores 0.

I compare all applications before advancing anyone. If the pool of 5 out of 5 scores is larger than 15, I rank by depth of regulated environment experience and scale of systems owned. The top 15 go forward. If fewer than 15 score 5 out of 5, all of them go forward.

The video is reviewed by me and the team together. We are not assessing your camera confidence. We are assessing whether your answer is specific, whether you owned what you are describing, and whether your response to a real production failure was sound.

I do not follow up to ask for clarification on an ambiguous CV. What is written is what is scored.

You will hear back from us regardless of outcome. That is a promise, not a pleasantry.

Hubert Warszta
Tech Recruiter | WhyHireWrong? |

Related Jobs

View all jobs

Senior MLOps Engineer

Machine Learning Engineer

Machine Learning Engineer (Manager)

On Senior Lead - Machine Learning Engineering

Senior Data Scientist

Lead Data Scientist

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many AI Tools Do You Need to Know to Get an AI Job?

If you are job hunting in AI right now it can feel like you are drowning in tools. Every week there is a new framework, a new “must-learn” platform or a new productivity app that everyone on LinkedIn seems to be using. The result is predictable: job seekers panic-learn a long list of tools without actually getting better at delivering outcomes. Here is the truth most hiring managers will quietly agree with. They do not hire you because you know 27 tools. They hire you because you can solve a problem, communicate trade-offs, ship something reliable and improve it with feedback. Tools matter, but only in service of outcomes. So how many AI tools do you actually need to know? For most AI job seekers: fewer than you think. You need a tight core toolkit plus a role-specific layer. Everything else is optional. This guide breaks it down clearly, gives you a simple framework to choose what to learn and shows you how to present your toolset on your CV, portfolio and interviews.

What Hiring Managers Look for First in AI Job Applications (UK Guide)

Hiring managers do not start by reading your CV line-by-line. They scan for signals. In AI roles especially, they are looking for proof that you can ship, learn fast, communicate clearly & work safely with data and systems. The best applications make those signals obvious in the first 10–20 seconds. This guide breaks down what hiring managers typically look for first in AI applications in the UK market, how to present it on your CV, LinkedIn & portfolio, and the most common reasons strong candidates get overlooked. Use it as a checklist to tighten your application before you click apply.

The Skills Gap in AI Jobs: What Universities Aren’t Teaching

Artificial intelligence is no longer a future concept. It is already reshaping how businesses operate, how decisions are made, and how entire industries compete. From finance and healthcare to retail, manufacturing, defence, and climate science, AI is embedded in critical systems across the UK economy. Yet despite unprecedented demand for AI talent, employers continue to report severe recruitment challenges. Vacancies remain open for months. Salaries rise year on year. Candidates with impressive academic credentials often fail technical interviews. At the heart of this disconnect lies a growing and uncomfortable truth: Universities are not fully preparing graduates for real-world AI jobs. This article explores the AI skills gap in depth—what is missing from many university programmes, why the gap persists, what employers actually want, and how jobseekers can bridge the divide to build a successful career in artificial intelligence.