Senior Machine Learning Engineer - AI & GPU Performance

synthesia.io
City of London
4 months ago
Applications closed

Related Jobs

View all jobs

Senior Machine Learning Engineer

Senior Machine Learning Engineer

Senior Machine Learning Engineer

Senior Machine Learning Engineer

Senior Machine Learning Engineer

Senior Machine Learning Engineer

Who are we?

From your everyday PowerPoint presentations to Hollywood movies, AI will transform the way we create and consume content.


Today, people want to watch and listen, not read — both at home and at work. If you’re reading this and nodding, check out our brand video.


Despite the clear preference for video, communication and knowledge sharing in the business environment are still dominated by text, largely because high‑quality video production remains complex and challenging to scale—until now….


Meet Synthesia

We're on amission to make video easy for everyone. Born in an AI lab, our AI video communications platform simplifies the entire video production process, making it easy for everyone, regardless of skill level, to create, collaborate, and share high‑quality videos. Whether it's for delivering essential training to employees and customers or marketing products and services, Synthesia enables large organizations to communicate and share knowledge through video quickly and efficiently. We’re trusted by leading brands such as Heineken, Zoom, Xerox, McDonald’s and more. Readstories from happy customers and what1,200+ people say on G2.


In February 2024, G2 named us as the fastest growing company in the world. Today, we're at a $2.1bn valuation and we recently raised our Series D. This brings our total funding to over $330M from top‑tier investors, including Accel, Nvidia, Kleiner Perkins, Google and top founders and operators including Stripe, Datadog, Miro, Webflow, and Facebook.


About the role

As a ML Performance Engineer in the AI & GPU Performance team you will contribute to the design and development of high performance solutions. You will join a team of 40+ Researchers and Engineers within the R&D Department working on cutting edge challenges in the Generative AI space, with a focus on creating highly realistic, emotional and life‑like Synthetic humans through text‑to‑video. Within the team you’ll have the opportunity to work on the applied side of our research efforts and directly impact our solutions that are used worldwide by over 60,000 businesses.


This is an opportunity to work for a company that is impacting businesses at a rapid pace across the globe.


What will you be doing?

As a ML Performance Engineer in the AI & GPU Performance team you will contribute to the design and development of high performance solutions. You will own one or more projects for computationally optimizing large‑scale model training and inference pipelines. By partnering with researchers and research teams you’ll identify high‑impact initiatives and push the boundaries of model performance. You will work on re‑implementing models in an efficient manner by using PyTorch and underlying technologies like CUDA/Triton, Torch compilation, etc.


This would include:



  • Evaluating, profiling and optimising compute resource usage (e.g., Hopper & Blackwell GPUs) for cost and time efficiency at training and inference times
  • Developing customised efficient solutions for inference pipelines (CUDA/Triton kernels) as well as introducing or enhancing tooling for achieving optimal computational performance (e.g. DL compilers, ONNX, TensorRT)
  • Driving the adoption of best practices for large‑model training, including checkpointing, gradient accumulation, and memory optimisation among others
  • Introducing or enhancing tooling for distributed training, performance monitoring, and logging (e.g., DeepSpeed, PyTorch Distributed)
  • Designing and implementing techniques for model parallelism, data parallelism, and mixed‑precision training
  • Keeping updated on the latest research in model compression (e.g., quantization, pruning) and advanced optimisation methods

Who are you?

  • You are an ML engineer passionate about high performance computing
  • You have a background in Computer Science / Engineering and 3+ years of industry experience. (PhD preferred)
  • You have worked on optimising large models for over 2 years
  • You have experience developing CUDA/Triton kernels and optimizing models with DL compilers (torch.compile)
  • You have great coding skills in Python and C++ and you care about writing clean, and efficient code
  • You have experience with optimising distributed systems and distributed tools like DDP, Deepspeed, Accelerate or similar
  • You have some experience in the video space (Diffusion models / GAN’s)
  • You are interested in doing research, trying new things and pushing the boundaries, going beyond what's already known

The good stuff…

  • Attractive compensation (salary + stock options + bonus)
  • Private Health Insurance in London
  • Hybrid work setting with an office in London
  • 25 days of annual leave + public holidays
  • Work in a great company culture with the option to join regular planning and socials at our hubs.
  • A generous referral scheme when you know people that are amazing for us
  • Strong opportunities for your career growth

Interested in building your career at Synthesia? Get future opportunities sent straight to your email.


By checking this box, I agree to allow Synthesia to retain my data for future opportunities for employment for up to 700 days after the conclusion of consideration of my current application for employment.


#J-18808-Ljbffr

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many AI Tools Do You Need to Know to Get an AI Job?

If you are job hunting in AI right now it can feel like you are drowning in tools. Every week there is a new framework, a new “must-learn” platform or a new productivity app that everyone on LinkedIn seems to be using. The result is predictable: job seekers panic-learn a long list of tools without actually getting better at delivering outcomes. Here is the truth most hiring managers will quietly agree with. They do not hire you because you know 27 tools. They hire you because you can solve a problem, communicate trade-offs, ship something reliable and improve it with feedback. Tools matter, but only in service of outcomes. So how many AI tools do you actually need to know? For most AI job seekers: fewer than you think. You need a tight core toolkit plus a role-specific layer. Everything else is optional. This guide breaks it down clearly, gives you a simple framework to choose what to learn and shows you how to present your toolset on your CV, portfolio and interviews.

What Hiring Managers Look for First in AI Job Applications (UK Guide)

Hiring managers do not start by reading your CV line-by-line. They scan for signals. In AI roles especially, they are looking for proof that you can ship, learn fast, communicate clearly & work safely with data and systems. The best applications make those signals obvious in the first 10–20 seconds. This guide breaks down what hiring managers typically look for first in AI applications in the UK market, how to present it on your CV, LinkedIn & portfolio, and the most common reasons strong candidates get overlooked. Use it as a checklist to tighten your application before you click apply.

The Skills Gap in AI Jobs: What Universities Aren’t Teaching

Artificial intelligence is no longer a future concept. It is already reshaping how businesses operate, how decisions are made, and how entire industries compete. From finance and healthcare to retail, manufacturing, defence, and climate science, AI is embedded in critical systems across the UK economy. Yet despite unprecedented demand for AI talent, employers continue to report severe recruitment challenges. Vacancies remain open for months. Salaries rise year on year. Candidates with impressive academic credentials often fail technical interviews. At the heart of this disconnect lies a growing and uncomfortable truth: Universities are not fully preparing graduates for real-world AI jobs. This article explores the AI skills gap in depth—what is missing from many university programmes, why the gap persists, what employers actually want, and how jobseekers can bridge the divide to build a successful career in artificial intelligence.