Senior Data Engineer

Wavendon
3 months ago
Applications closed

Related Jobs

View all jobs

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

At Unisys, we are dedicated to building high-performance, secure digital solutions that drive innovation and transformation for our clients. With a legacy of over a century in technology leadership, we empower businesses and governments around the world to achieve new levels of efficiency, security, and customer satisfaction. Our team thrives on solving complex challenges with cutting-edge technologies, and we are proud to be a trusted partner to some of the world’s most respected organizations.

We are seeking an experienced, hands-on Senior Engineer with specialized skills working on developing applications and a flair for Generative AI and AI/LM technologies. The engineer will be responsible for:

• Quickly converting an idea to a demonstration

• Lead the design and development of Generative AI-based applications.

• Working within in a Scrum team to develop rapid prototypes.

• Evaluate, select, and integrate AI tools and frameworks that are essential for generative AI development.

• Stay up to date with the latest advancements in generative AI and contribute to our research efforts in this field.

• Learning, evaluating and leveraging a variety of AI tools and technologies.

• Being a self-starter and able to work independently, as well as within a team, to deliver a result.

• Provide guidance and mentorship to junior AI engineers, fostering a culture of continuous learning.

• Demonstrate passion for quality and productivity by use of efficient development techniques, standards and guidelines.

• Design, train, fine-tune, and evaluate ML and NLP models (including small/medium language models) using robust experimentation practices, clear success metrics, and offline/online evaluation.

• Build and maintain data pipelines for ML (ingestion, validation, feature engineering, labeling) with reproducibility, lineage, and governance.

• Implement retrieval-augmented generation (RAG) and hybrid search, including embeddings, vector stores, and re-ranking to improve quality and guardrails.

• Optimize LLM/SLM inference (quantization, distillation, batching, caching) for latency, throughput, and cost; select and tune serving stacks (e.g., vLLM/TGI/Triton) and GPU/CPU targets.

• Operationalize models with MLOps/AIOps best practices: CI/CD for models, model/version registries, feature stores, automated testing, canary/blue–green rollouts, A/B tests, and rollback strategies.

• Deploy, scale, and monitor models on Kubernetes (Helm, Rancher, AKS, K3S) and cloud AI platforms; instrument for observability (metrics, logs, traces), data/label drift, bias, hallucination and safety events.

• Implement security, compliance, and privacy-by-design for AI systems (PII handling, policy enforcement, content moderation, prompt/response safety, secret management).

• Partner with Data Engineering and Platform teams to ensure capacity planning, cost controls, and reliability (SLOs/SLIs) for model serving in production.

You will be successful in this role if you have:

• Must have proficient skills working on Large Language Models (LLMs).

• Proficiency in evaluating, deploying and using on premise open source technology stacks.

• Proficiency in Azure Cognitive Services, which includes Azure Language Service, Azure Text Analytics, Azure Speech Service, and other AI-related offerings.

• A good understanding of OpenAI's GPT models and how it integrates with Azure services. This includes knowledge of GPT's capabilities, limitations, and available features.

• Strong programming skills in Python for working with Azure services and data manipulation.

• Good understanding of databases e.g. Postgres

• Proficient knowledge of back-end programming languages like NodeJS, Python and/or Golang

• Proficient knowledge on one of front-end technologies like React or Angular

• Hands on technical Skills related to Backend development.

• Experience with AWS and GCP is an added advantage.

• Demonstrated Cluster Management knowledge and experience using platforms including – Kubernetes, Rancher, Helm, Docker

• Familiarity with Figma for UI Mockups

• Cloud certification at a developer or equivalent level is an added advantage.

• Azure AI Certification or equivalent is an added advantage.

• Familiarity with Data engineering and Machine learning models

• Excellent written and verbal communication skills, flexible and good attitude

• Experience working with open-source projects, third-party libraries, SDKs and APIs

• Practical experience training/fine-tuning SLMs/LLMs (e.g., Llama, Mistral) using techniques such as LoRA/QLoRA, PEFT, and prompt-tuning; knowledge of evaluation frameworks and benchmark design.

• Strong grasp of ML fundamentals (supervised/unsupervised learning, bias/variance, cross-validation), NLP techniques, and vector embeddings; hands-on with PyTorch/TensorFlow and scikit-learn.

• Experience with LLM/RAG tooling and NLP stacks (Hugging Face Transformers/Datasets, LangChain or LlamaIndex) and vector databases (pgvector/Postgres, Pinecone, Milvus, or Azure AI Search).

• Proficiency with experiment tracking and model registries (MLflow, Weights & Biases) and feature stores; comfort with disciplined experiment design and reproducibility.

• Model serving and optimization expertise: vLLM, ollama, Text Generation Inference, Triton, ONNX Runtime; quantization (bitsandbytes/INT8/FP8, GPTQ/AWQ), distillation, and caching/batching strategies.

• MLOps/AIOps experience building CI/CD for ML (Jenkins/Azure DevOps), automated testing of data and models, canary releases, model rollbacks, and online A/B experimentation.

• Production monitoring and observability (Prometheus/Grafana/OpenTelemetry/ELK), data and concept drift detection, safety/hallucination monitoring, and feedback loops for continual learning.

• Data engineering proficiency with Spark/Databricks and/or Kafka for streaming features, plus robust data quality/lineage tooling.

• Understanding of GPU/accelerator concepts (CUDA, NCCL, memory/throughput tradeoffs) and capacity planning for cost-effective scaling.

• Knowledge of AI safety, security, and compliance practices (RBAC, secrets management, PII handling, red-teaming, prompt injection defenses, content moderation).

• Experience with managed AI platforms (Azure AI Studio/Model Catalog, Azure ML, SageMaker, Vertex AI) and multi-cloud deployments is a plus

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Data Science Tools Do You Need to Know to Get a Data Science Job?

If you’re trying to break into data science — or progress your career — it can feel like you are drowning in names: Python, R, TensorFlow, PyTorch, SQL, Spark, AWS, Scikit-learn, Jupyter, Tableau, Power BI…the list just keeps going. With every job advert listing a different combination of tools, many applicants fall into a trap: they try to learn everything. The result? Long tool lists that sound impressive — but little depth to back them up. Here’s the straight-talk version most hiring managers won’t explicitly tell you: 👉 You don’t need to know every data science tool to get hired. 👉 You need to know the right ones — deeply — and know how to use them to solve real problems. Tools matter, but only in service of outcomes. So how many data science tools do you actually need to know to get a job? For most job seekers, the answer is not “27” — it’s more like 8–12, thoughtfully chosen and well understood. This guide explains what employers really value, which tools are core, which are role-specific, and how to focus your toolbox so your CV and interviews shine.

What Hiring Managers Look for First in Data Science Job Applications (UK Guide)

If you’re applying for data science roles in the UK, it’s crucial to understand what hiring managers focus on before they dive into your full CV. In competitive markets, recruiters and hiring managers often make their first decisions in the first 10–20 seconds of scanning an application — and in data science, there are specific signals they look for first. Data science isn’t just about coding or statistics — it’s about producing insights, shipping models, collaborating with teams, and solving real business problems. This guide helps you understand exactly what hiring managers look for first in data science applications — and how to structure your CV, portfolio and cover letter so you leap to the top of the shortlist.

The Skills Gap in Data Science Jobs: What Universities Aren’t Teaching

Data science has become one of the most visible and sought-after careers in the UK technology market. From financial services and retail to healthcare, media, government and sport, organisations increasingly rely on data scientists to extract insight, guide decisions and build predictive models. Universities have responded quickly. Degrees in data science, analytics and artificial intelligence have expanded rapidly, and many computer science courses now include data-focused pathways. And yet, despite the volume of graduates entering the market, employers across the UK consistently report the same problem: Many data science candidates are not job-ready. Vacancies remain open. Hiring processes drag on. Candidates with impressive academic backgrounds fail interviews or struggle once hired. The issue is not intelligence or effort. It is a persistent skills gap between university education and real-world data science roles. This article explores that gap in depth: what universities teach well, what they often miss, why the gap exists, what employers actually want, and how jobseekers can bridge the divide to build successful careers in data science.