Lead Data Engineer

Kanzlei Ganz Gärtner Lindberg Slania
Manchester
1 month ago
Applications closed

Related Jobs

View all jobs

Lead Data Engineer

Lead Data Engineer - Azure Synapse

Lead Data Engineer - Hadoop - Spark - Python

Lead Data/Head of Data Engineer

Lead Data Engineer

Lead Data Engineer

Lead Data Engineer

Join Lead Data Engineer role at Kanzlei Ganz Gärtner Lindberg Slania.

About the Role

We’re looking for a Lead Data Engineer to join our Data Engineering and Analytics practice.

In This Role
  • Lead the design, development, management and optimisation of data pipelines to ensure efficient data flows, recognising and sharing opportunities to reuse data flows where possible.
  • Coordinate teams and set best practices and standards for data engineering principles.
  • Champion data engineering across projects and clients.
Responsibilities
  • Lead by example, holding responsibilities for team culture and project impact.
  • Be accountable for the strategic direction, delivery and growth of our work.
  • Lead teams, strands of work and outcomes, owning commercial responsibilities.
  • Hold and manage uncertainty and ambiguity on behalf of clients and teams.
  • Ensure teams and projects are inclusive through how you lead and manage others.
  • Effectively own and hold the story of our work, measuring progress against client goals and our DT missions.
  • Work with our teams to influence and own how we deliver more value to clients, working with time and budget constraints.
  • Strategically plan the overall project and apply methods and approaches.
  • Demonstrably share work with wider audiences.
  • Elevate ideas through writing, speaking and presenting.
Dimensions
  • Headcount: Typically leads a multidisciplinary team or multiple workstreams (team size 5‑15).
  • Resource complexity: Provides leadership across multiple workstreams or technical domains, overseeing junior leads or specialists.
  • Problem‑solving responsibility: Solves highly complex problems, balancing technical, user, business, and operational needs.
  • Change management: Leads or co‑leads significant change initiatives, managing stakeholder expectations and embedding sustainable ways of working.
  • Internal/External interactions: Acts as a trusted partner to client and internal stakeholders, leading workshops and presentations.
  • Strategic timeframe: Works across mid‑to‑long term delivery cycles (6‑12 months).
About You – Essential
  • Proven experience in data engineering, data integration and data modelling.
  • Expertise with cloud platforms (e.g., AWS, Azure, GCP).
  • Expertise with modern cloud data platforms (e.g., Microsoft Fabric, Databricks).
  • Expertise with multiple data analytics tools (e.g., Power BI).
  • Deep understanding of data warehousing concepts, ETL/ELT pipelines and dimensional modelling.
  • Proficiency in programming languages (Python/PySpark, SQL).
  • Experience in data pipeline orchestration (e.g., Airflow, Data Factory).
  • Familiarity with DevOps and CI/CD practices (Git, Azure DevOps).
  • Ability to communicate technical concepts to technical and non‑technical audiences.
  • Proven experience delivering complex projects in a fast‑paced environment with tight deadlines.
About You – Desirable
  • Advanced knowledge of data governance, data standards and best practices.
  • Experience in a consultancy environment, demonstrating flexibility and adaptability.
  • Experience defining and enforcing data engineering standards, patterns and reusable frameworks.
  • Professional certifications (e.g., Microsoft Azure Data Engineer, AWS Data Analytics, Databricks Certified Professional Data Engineer).
Skills
  • Design, build and test complex or large‑scale data products.
  • Build and lead teams to deliver data integration services and reusable pipelines meeting performance, quality and scalability standards.
  • Collaborate with architects to align solutions with enterprise data strategy.
  • Work with data analysts, engineers and data science/AI specialists to design and deliver products effectively.
  • Understand data cleansing and preparation, implementing reusable processes and checks.
  • Optimize data pipelines and queries for performance and cost efficiency.
  • Define system integration testing conditions for complex data products, support others, analyze and report test activities.
  • Develop and maintain complex data models (conceptual, logical, physical).
  • Strong skills in data governance and metadata management.
  • Experience with CI/CD pipelines, version control, and infrastructure-as-code (Git, Azure DevOps).
  • Strong stakeholder communication, translating technical concepts into business terms.
  • Mentor junior engineers and build a high‑performing data engineering culture.
Behaviours and PACT Values
  • Purpose – values‑driven, putting client needs first.
  • Accountability – deliver on time and under budget, promoting quality and client experience.
  • Craft – balance priorities, navigate ambiguity, set technical direction.
  • Togetherness – collaborate effectively, build strong relationships.
About Us

We’re a purpose‑driven organisation supporting organisations to build a better future for people, places and the planet. We combine experience in the public, private and third sectors with expertise in human‑centred design, data, experience and technology to create sustainable solutions for an ever‑evolving world.

Benefits
  • 30 days holiday + bank holidays
  • 2 volunteer days for causes you care about
  • Maternity/paternity – 6 months maternity leave, 3 months paternity leave
  • Life assurance
  • Employer pension contribution of 5%
  • Health cash plan
  • Personal learning and development budget
  • Employee Assistance Programme
  • Access to equity via a Share Incentive Plan
  • Green incentive programmes (Electric Vehicle Leasing, Cycle to Work Scheme)
  • Financial advice
  • Health assessments


#J-18808-Ljbffr

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Data Science Tools Do You Need to Know to Get a Data Science Job?

If you’re trying to break into data science — or progress your career — it can feel like you are drowning in names: Python, R, TensorFlow, PyTorch, SQL, Spark, AWS, Scikit-learn, Jupyter, Tableau, Power BI…the list just keeps going. With every job advert listing a different combination of tools, many applicants fall into a trap: they try to learn everything. The result? Long tool lists that sound impressive — but little depth to back them up. Here’s the straight-talk version most hiring managers won’t explicitly tell you: 👉 You don’t need to know every data science tool to get hired. 👉 You need to know the right ones — deeply — and know how to use them to solve real problems. Tools matter, but only in service of outcomes. So how many data science tools do you actually need to know to get a job? For most job seekers, the answer is not “27” — it’s more like 8–12, thoughtfully chosen and well understood. This guide explains what employers really value, which tools are core, which are role-specific, and how to focus your toolbox so your CV and interviews shine.

What Hiring Managers Look for First in Data Science Job Applications (UK Guide)

If you’re applying for data science roles in the UK, it’s crucial to understand what hiring managers focus on before they dive into your full CV. In competitive markets, recruiters and hiring managers often make their first decisions in the first 10–20 seconds of scanning an application — and in data science, there are specific signals they look for first. Data science isn’t just about coding or statistics — it’s about producing insights, shipping models, collaborating with teams, and solving real business problems. This guide helps you understand exactly what hiring managers look for first in data science applications — and how to structure your CV, portfolio and cover letter so you leap to the top of the shortlist.

The Skills Gap in Data Science Jobs: What Universities Aren’t Teaching

Data science has become one of the most visible and sought-after careers in the UK technology market. From financial services and retail to healthcare, media, government and sport, organisations increasingly rely on data scientists to extract insight, guide decisions and build predictive models. Universities have responded quickly. Degrees in data science, analytics and artificial intelligence have expanded rapidly, and many computer science courses now include data-focused pathways. And yet, despite the volume of graduates entering the market, employers across the UK consistently report the same problem: Many data science candidates are not job-ready. Vacancies remain open. Hiring processes drag on. Candidates with impressive academic backgrounds fail interviews or struggle once hired. The issue is not intelligence or effort. It is a persistent skills gap between university education and real-world data science roles. This article explores that gap in depth: what universities teach well, what they often miss, why the gap exists, what employers actually want, and how jobseekers can bridge the divide to build successful careers in data science.