Senior Data Engineer

SOLVACE
Glasgow
3 days ago
Create job alert

The Opportunity - (Fully Remote, Perm role)


This is a greenfield data platform build. Solvace is in the middle of a major AI transformation — a multi‐agent copilot platform (KAI) is already live in production, serving manufacturing clients globally. The AI roadmap is ambitious (agentic AI capabilities, multi‐agent orchestration, voice interfaces), and the next wave of capabilities depends on a strong data engineering foundation. A Databricks lakehouse POC is being evaluated (Unity Catalog, Delta Sharing, PrivateLink to RDS), and the team is assessing the right long‐term analytics platform strategy. This role builds the entire data engineering function from scratch — defining the architecture, tooling, and patterns that power the next generation of AI capabilities.


This is an entirely forward‐looking role. While there is an active platform modernisation underway (the engineering team is migrating a .NET platform from SQL Server to PostgreSQL and containerising for Kubernetes), this role’s primary focus is building new data pipelines and analytics infrastructure, not maintaining legacy systems. That said, collaboration with the platform engineering team is important, as the data pipelines need to ingest from both the existing SQL Server estate and the new PostgreSQL databases.


Core Technical Requirements

• Python — primary language for data pipelines, scripting, and integration with the AI stack. This is a Python‐first team

• SQL (SQL Server + PostgreSQL) — deep experience required across both engines. The platform runs a large‐scale multi‐tenant database estate with per‐client schemas (local, global, corporate). The new platform targets Aurora PostgreSQL, with several modules already migrated. Pipelines must bridge both.

• Databricks / Apache Spark — a lakehouse POC is being evaluated with Unity Catalog and Delta Sharing via AWS PrivateLink. Experience with Databricks or equivalent analytics platforms is important for assessing and scaling the right approach

• Apache Kafka — event streaming experience is essential. The platform is evaluating Kafka as the strategic backbone for ordered event logs, consumer group replay, and guaranteed partition ordering as the microservices fleet scales

• ETL/ELT pipeline design — building and maintaining pipelines into a medallion architecture (bronze/silver/gold layers) on Databricks or equivalent analytics platform

• Delta Lake — understanding of lakehouse table formats, time travel, and ACID transaction patterns

• AWS data services — RDS, S3, Lambda, Glue or equivalent. The platform runs entirely on AWS • Linux / CLI proficiency — must be comfortable working



AI-Assisted Development Methodology ->


Solvace is transitioning towards AI‐assisted development as a core engineering practice. Candidates should demonstrate:

• Hands‐on experience with AI coding tools — Claude Code, OpenAI Codex, GitHub Copilot, Cursor, or similar. We’re looking for engineers who have integrated these tools into their professional workflow, not just experimented casually

• Spec‐driven development — ability to write clear technical specifications that can be used to drive both human and AI‐assisted implementation, with strong evaluation criteria and test coverage

• Portfolio evidence — professional projects or side projects that demonstrate AI‐assisted development practices. Contributions to or experimentation with emerging projects like OpenClaw are a strong signal

• Testing and evaluation rigour — experience building robust test suites, automated quality gates, and evaluation frameworks that ensure AI‐assisted code meets production standards


Nice-to-Have ->

• Apache Flink — for real‐time stream processing as the platform moves from batch to streaming analytics

• Apache Iceberg — experience with open table formats alongside or as an alternative to Delta Lake

• Go or Rust — valued as evidence of strong backend engineering and systems thinking, even though Python is the primary language

• Multi‐tenant data architectures — the platform runs per‐client schemas with tenant isolation requirements. Understanding tenant isolation patterns is a significant advantage

• Terraform / IaC — infrastructure is managed via Terraform; the Databricks POC has been codified in Terraform

• dbt or similar transformation framework

• Manufacturing / industrial data domain experience (sensor data, quality metrics, OEE, SPC)

• CDC (Change Data Capture) — the platform is building CDC pipelines for the SQL Server to PostgreSQL transition

• Weaviate or other vector database experience — currently used for semantic search in the AI agent layer


Why Join?

• Greenfield data platform — the entire data engineering practice needs to be built from scratch. A lakehouse POC is being evaluated but the production architecture is theirs to define — patterns, tooling, and platform decisions from day one

• Direct AI impact — every pipeline they build directly unblocks AI agent capabilities. The KAI copilot is live in production and evolving towards agentic AI — autonomous agents that take actions, orchestrate workflows, and operate across systems. This role provides the data foundation that makes agentic capabilities possible

• Event streaming architecture — opportunity to design and implement a Kafka‐based event backbone, shaping the platform’s real‐time data architecture

• Manufacturing AI — opportunity to work at the intersection of industrial operations and generative AI, a domain with massive untapped potential. Rich, real‐world data: quality inspections, KPIs, maintenance records, safety observations, process parameters across global manufacturing sites

• Small team, high autonomy — the Innovation Hub operates with startup‐level autonomy inside a funded, growing company

• Leadership trajectory — the data function is critical to the platform and will grow. As an early hire, there is an open career path that will grow with the function — earned, not guaranteed

Related Jobs

View all jobs

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

New Data Science Employers to Watch in 2026: UK and International Companies Leading Analytics and AI Innovation

Data science has emerged as one of the most transformative forces across industries, turning raw information into actionable insights, predictive models, and AI-powered solutions. In 2026, the UK is witnessing a surge in organisations where data science is not just a support function but the core of their products and services. For professionals exploring opportunities on www.DataScience-Jobs.co.uk , identifying these employers early can provide a competitive advantage in a market with high demand for advanced analytics and machine learning expertise. This article highlights new and high-growth data science employers to watch in 2026, focusing on UK startups, scale-ups, and global firms expanding their data science operations locally. All of the companies included have recently raised investment, won high-profile contracts, or significantly scaled their analytics teams.

How Many Data Science Tools Do You Need to Know to Get a Data Science Job?

If you’re trying to break into data science — or progress your career — it can feel like you are drowning in names: Python, R, TensorFlow, PyTorch, SQL, Spark, AWS, Scikit-learn, Jupyter, Tableau, Power BI…the list just keeps going. With every job advert listing a different combination of tools, many applicants fall into a trap: they try to learn everything. The result? Long tool lists that sound impressive — but little depth to back them up. Here’s the straight-talk version most hiring managers won’t explicitly tell you: 👉 You don’t need to know every data science tool to get hired. 👉 You need to know the right ones — deeply — and know how to use them to solve real problems. Tools matter, but only in service of outcomes. So how many data science tools do you actually need to know to get a job? For most job seekers, the answer is not “27” — it’s more like 8–12, thoughtfully chosen and well understood. This guide explains what employers really value, which tools are core, which are role-specific, and how to focus your toolbox so your CV and interviews shine.

What Hiring Managers Look for First in Data Science Job Applications (UK Guide)

If you’re applying for data science roles in the UK, it’s crucial to understand what hiring managers focus on before they dive into your full CV. In competitive markets, recruiters and hiring managers often make their first decisions in the first 10–20 seconds of scanning an application — and in data science, there are specific signals they look for first. Data science isn’t just about coding or statistics — it’s about producing insights, shipping models, collaborating with teams, and solving real business problems. This guide helps you understand exactly what hiring managers look for first in data science applications — and how to structure your CV, portfolio and cover letter so you leap to the top of the shortlist.