Be at the heart of actionFly remote-controlled drones into enemy territory to gather vital information.

Apply Now

Data Scientist (ML, Speech, NLP & Multimodal Expertise) | Manchester

Transperfect Gaming Solutions
Manchester
1 week ago
Create job alert
Overview

We are looking to hire a Data Scientist with strong expertise in machine learning, speech and language processing, and multimodal systems. This role is essential to driving our product roadmap forward, particularly in building out our core machine learning systems and developing next-generation speech technologies. The ideal candidate will be capable of working independently while effectively collaborating with cross-functional teams, and will be curious, experimental, and communicative.

Key Responsibilities
  • Create maintainable, elegant code and high-quality data products that are modeled, well-documented, and simple to use.
  • Build, maintain, and improve the infrastructure to extract, transform, and load data from a variety of sources using SQL, Azure, GCP and AWS technologies.
  • Perform statistical analysis of training datasets to identify biases, quality issues, and coverage gaps.
  • Implement automated evaluation pipelines that scale across multiple models and tasks.
  • Create interactive dashboards and visualization tools for model performance analysis.
Additional Responsibilities
  • Design and implement robust data ingestion pipelines for massive-scale text and speech corpora including automated data preprocessing and cleaning pipelines.
  • Create data validation frameworks and monitoring systems for dataset quality.
  • Develop sampling strategies for balanced and representative training data.
  • Implement comprehensive experiment tracking and hyperparameter optimization frameworks.
  • Conduct statistical analysis of training dynamics and convergence patterns.
  • Design A/B testing frameworks for comparing different training approaches.
  • Create automated model selection pipelines based on multiple evaluation criteria.
  • Develop cost-benefit analyses for different training configurations.
  • Design comprehensive benchmark suites with statistical significance testing.
  • Develop fairness metrics and bias detection systems.
  • Build real-time monitoring systems for model performance in production.
  • Implement feature drift detection and data quality monitoring.
  • Design feedback loops to capture user interactions and model effectiveness.
  • Create automated retraining pipelines based on performance degradation signals.
  • Develop business metrics and ROI analysis for model deployments.
Required Skills, Experience and QualificationsProgramming & Software Engineering
  • Python (Expert Level): Advanced proficiency in scientific computing stack (NumPy, Pandas, SciPy, Scikit-learn).
  • Version Control: Git workflows, collaborative development, and code review processes.
  • Software Engineering Practices: Testing frameworks, CI/CD pipelines, and production-quality code development.
Machine Learning and Language Model Expertise
  • Traditional Machine Learning and Deep Learning Knowledge: Proficiency in classical ML algorithms (Naive Bayes, SVM, Random Forest, etc.) and Deep Learning architectures.
  • Understanding of Transformer Architecture: Attention mechanisms, positional encoding, and scaling laws.
  • Training Pipeline Knowledge: Data preprocessing for large corpora, tokenization strategies, and distributed training concepts.
  • Evaluation Frameworks: Experience with standard NLP benchmarks (GLUE, SuperGLUE, etc.) and custom evaluation design.
  • Fine-tuning Techniques: Understanding of PEFT methods, instruction tuning, and alignment techniques.
  • Model Deployment: Knowledge of model optimization, quantization, and serving infrastructure for large models.
Collaboration & Adaptability
  • Strong communication skills are a must
  • Self-reliant but knows when to ask for help
  • Comfortable working in an environment where conventional development practices may not always apply
  • PBIs (Product Backlog Items) may not be highly detailed
  • Experimentation will be necessary
  • Ability to identify what's important in completing a task or partial task and explain/justify their approach
  • Can effectively communicate ideas and strategies
  • Proactive and takes initiative rather than waiting for PBIs to be assigned when circumstances call for it
  • Strong interest in AI and its possibilities, a genuine passion for certain areas can provide that extra spark
  • Curious and open to experimenting with technologies or languages outside their comfort zone
Mindset & Work Approach
  • Takes ownership when things don't go as planned
  • Capable of working from high-level explanations and general guidance on implementations and final outcomes
  • Continuous, clear communication is crucial, detailed step-by-step instructions may not always be available
  • Self-starter, self-motivated, and proactive in problem-solving
  • Enjoys exploring and testing different approaches, even in unfamiliar programming languages
Additional Skills, Experience and QualificationsMachine Learning & Deep Learning
  • Framework Proficiency: Scikit-learn, XGBoost, PyTorch (preferred) or TensorFlow for model implementation and experimentation.
  • MLOps Expertise: Model versioning, experiment tracking, model monitoring (MLflow, Weights & Biases), data monitoring and validation (Great Expectations, Prometheus, Grafana), and automated ML pipelines (GitHub CI/CD, Jenkins, CircleCI, GitLab etc.).
  • Statistical Modeling: Hypothesis testing, experimental design, causal inference, and Bayesian statistics.
  • Model Evaluation: Cross-validation strategies, bias-variance analysis, and performance metric design.
  • Feature Engineering: Advanced techniques for text, time-series, and multimodal data.
Data Engineering & Infrastructure
  • Big Data Technologies: Spark (PySpark), Hadoop ecosystem, and distributed computing frameworks (DDP, TP, FSDP).
  • Cloud Platforms: AWS (SageMaker, S3, EMR), GCP (Vertex AI, BigQuery), or Azure ML.
  • Database Systems: NoSQL databases (MongoDB, Elasticsearch), graph databases (Neo4j), and vector databases (Pinecone, Milvus, ChromaDB, FAISS etc.).
  • Data Pipeline Tools: Airflow, Prefect, or similar orchestration frameworks.
  • Containerization: Docker, Kubernetes for scalable model deployment


#J-18808-Ljbffr

Related Jobs

View all jobs

Data Scientist

Data Scientist

Data Scientist

Data Scientist

Data Scientist

Data Scientist - Remote

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

Why AI Careers in the UK Are Becoming More Multidisciplinary

Artificial intelligence is no longer a single-discipline pursuit. In the UK, employers increasingly want talent that can code and communicate, model and manage risk, experiment and empathise. That shift is reshaping job descriptions, training pathways & career progression. AI is touching regulated sectors, sensitive user journeys & public services — so the work now sits at the crossroads of computer science, law, ethics, psychology, linguistics & design. This isn’t a buzzword-driven change. It’s happening because real systems are deployed in the wild where people have rights, needs, habits & constraints. As models move from lab demos to products that diagnose, advise, detect fraud, personalise education or generate media, teams must align performance with accountability, safety & usability. The UK’s maturing AI ecosystem — from startups to FTSE 100s, consultancies, the public sector & universities — is responding by hiring multidisciplinary teams who can anticipate social impact as confidently as they ship features. Below, we unpack the forces behind this change, spotlight five disciplines now fused with AI roles, show what it means for UK job-seekers & employers, and map practical steps to future-proof your CV.

AI Team Structures Explained: Who Does What in a Modern AI Department

Artificial Intelligence (AI) and Machine Learning (ML) are no longer confined to research labs and tech giants. In the UK, organisations from healthcare and finance to retail and logistics are adopting AI to solve problems, automate processes, and create new products. With this growth comes the need for well-structured teams. But what does an AI department actually look like? Who does what? And how do all the moving parts come together to deliver business value? In this guide, we’ll explain modern AI team structures, break down the responsibilities of each role, explore how teams differ in startups versus enterprises, and highlight what UK employers are looking for. Whether you’re an applicant or an employer, this article will help you understand the anatomy of a successful AI department.

Why the UK Could Be the World’s Next AI Jobs Hub

Artificial Intelligence (AI) has rapidly moved from research labs into boardrooms, classrooms, hospitals, and homes. It is already reshaping economies and transforming industries at a scale comparable to the industrial revolution or the rise of the internet. Around the world, countries are competing fiercely to lead in AI innovation and reap its economic, social, and strategic benefits. The United Kingdom is uniquely positioned in this race. With a rich heritage in computing, world-class universities, forward-thinking government policy, and a growing ecosystem of startups and enterprises, the UK has many of the elements needed to become the world’s next AI hub. Yet competition is intense, particularly from the United States and China. Success will depend on how effectively the UK can scale its strengths, close its gaps, and seize opportunities in the years ahead. This article explores why the UK could be the world’s next global hub for artificial intelligence, what challenges it must overcome, and what this means for businesses, researchers, and job seekers.