Senior DataOps / DevOps Engineer

hackajob
Newcastle upon Tyne
4 days ago
Create job alert

hackajob is collaborating with Sage to connect them with exceptional professionals for this role. We’re looking for a Senior DataOps / DevOps Engineer to design, build, and operate the reliability layer underpinning Sage’s core data platforms, including large-scale batch and streaming data systems. In this role, you’ll own the observability, monitoring, and operational resilience of cloud native data infrastructure and streaming pipelines, ensuring that data flows—whether event driven or batch—are performant, reliable, and predictable in production. This is a hybrid position requiring 3 days per week in our Newcastle office.


First 90 Days

  • 30 days: Get familiar with Sage’s data platform architecture, including batch and streaming pipelines, cloud infrastructure, and existing operational tooling. Understand current monitoring, alerting, logging, and incident response practices, along with data reliability SLAs, failure modes, and engineering standards.
  • 60 days: Begin actively improving observability across key data systems, including dashboards, alerts, and pipeline health checks. Contribute to the operation and reliability of batch and streaming workloads, applying Infrastructure as Code, incident learnings, and DataOps best practices.
  • 90 days: Own major aspects of the data platform’s operational reliability and observability strategy. Drive improvements in alert quality, system resilience, pipeline reliability, and operational maturity. Mentor team members on DataOps and DevOps practices, and help shape how data platforms are built and operated going forward.

Meet the Team

You’ll work alongside data engineers, AI specialists, product managers, and designers in a highly collaborative environment. The team focuses on building scalable internal platforms that power data-driven decision making and AI-enabled products across Sage.


How success will be measured

  • Delivery of reliable, scalable automation and operational capabilities across data ingestion, processing, and platform services.
  • Measurable improvements in platform observability, including clear dashboards and actionable alerts tied to data SLAs such as freshness, latency, and availability.
  • Reduction in operational toil through Infrastructure as Code, repeatable deployments, and improved self-service onboarding for engineering teams.
  • Improved incident response outcomes, including faster detection, faster recovery, and fewer recurring issues through effective post-incident followups.
  • Strong operational quality across environments, with platforms operating securely, predictably, and in line with governance and compliance requirements.
  • Increased visibility into system health across batch and streaming data pipelines.

Skills you’ll gain

  • Deep expertise operating a modern Product Data Platform / Data Hub supporting both batch and streaming workloads.
  • Hands-on experience with streaming and distributed data processing systems and their operational characteristics.
  • Strong exposure to observability engineering for data systems, including metrics, logs, traces, and pipeline health monitoring.
  • Experience shaping platform reliability standards, including alerting strategies, run books, and on call readiness.
  • Practical cloud infrastructure ownership across storage, compute, and analytics layers used by large scale data platforms.

Snapshot of your day to day

  • You’ll design and operate monitoring and alerting that provides realtime visibility into pipeline health, SLA breaches, and platform behaviour.
  • You’ll improve the reliability of batch and streaming data ingestion and processing workloads, focusing on failure recovery and operational robustness.
  • You’ll build and maintain cloud infrastructure and deployment automation to keep environments consistent, secure, and repeatable.
  • You’ll work closely with data engineering and product teams to improve platform onboarding and reduce the effort required to adopt shared data capabilities.
  • You’ll help strengthen governance, compliance, and auditability by improving observability, documentation, and operational controls across the platform.

Must have skills

  • Strong experience as a DataOps, DevOps, or Platform Engineer supporting production data systems.
  • Proven expertise in observability tooling, including monitoring, logging, alerting, dashboards, and operating distributed systems in production.
  • Solid understanding of streaming and event-driven data pipelines and their common failure modes (e.g. lag, back pressure, replay).
  • Strong cloud infrastructure experience (AWS preferred), including networking, compute, storage, and managed services.
  • Hands‑on experience with Infrastructure as Code and CI/CD practices for platform and data services.
  • Ability to work across ingestion, processing, and storage layers while collaborating effectively with multiple engineering teams.
  • Excellent communication and collaboration skills in English.

Nice to have skills

  • Experience operating data platforms built on technologies such as Snowflake and S3‑based data lake patterns.
  • Familiarity with distributed processing and streaming ecosystems such as Kafka or Flink.
  • Experience implementing data pipeline health monitoring beyond infrastructure metrics (e.g. freshness, completeness, anomaly detection).
  • Experience supporting multi‑team internal platforms with a “platform as a product” mindset.


#J-18808-Ljbffr

Related Jobs

View all jobs

Senior DataOps/DevOps Engineer - Data Platform Reliability

Senior DataOps Engineer, AWS Cloud Data Platform Reliability

Hybrid Senior DataOps Engineer (AWS & Observability)

Senior Azure DataOps & DevOps Engineer for Databricks

Senior DataOps Engineer: Scale and observe data platforms

Data Engineer - DataOps

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many AI Tools Do You Need to Know to Get an AI Job?

If you are job hunting in AI right now it can feel like you are drowning in tools. Every week there is a new framework, a new “must-learn” platform or a new productivity app that everyone on LinkedIn seems to be using. The result is predictable: job seekers panic-learn a long list of tools without actually getting better at delivering outcomes. Here is the truth most hiring managers will quietly agree with. They do not hire you because you know 27 tools. They hire you because you can solve a problem, communicate trade-offs, ship something reliable and improve it with feedback. Tools matter, but only in service of outcomes. So how many AI tools do you actually need to know? For most AI job seekers: fewer than you think. You need a tight core toolkit plus a role-specific layer. Everything else is optional. This guide breaks it down clearly, gives you a simple framework to choose what to learn and shows you how to present your toolset on your CV, portfolio and interviews.

What Hiring Managers Look for First in AI Job Applications (UK Guide)

Hiring managers do not start by reading your CV line-by-line. They scan for signals. In AI roles especially, they are looking for proof that you can ship, learn fast, communicate clearly & work safely with data and systems. The best applications make those signals obvious in the first 10–20 seconds. This guide breaks down what hiring managers typically look for first in AI applications in the UK market, how to present it on your CV, LinkedIn & portfolio, and the most common reasons strong candidates get overlooked. Use it as a checklist to tighten your application before you click apply.

The Skills Gap in AI Jobs: What Universities Aren’t Teaching

Artificial intelligence is no longer a future concept. It is already reshaping how businesses operate, how decisions are made, and how entire industries compete. From finance and healthcare to retail, manufacturing, defence, and climate science, AI is embedded in critical systems across the UK economy. Yet despite unprecedented demand for AI talent, employers continue to report severe recruitment challenges. Vacancies remain open for months. Salaries rise year on year. Candidates with impressive academic credentials often fail technical interviews. At the heart of this disconnect lies a growing and uncomfortable truth: Universities are not fully preparing graduates for real-world AI jobs. This article explores the AI skills gap in depth—what is missing from many university programmes, why the gap persists, what employers actually want, and how jobseekers can bridge the divide to build a successful career in artificial intelligence.