Data Engineer – GCP/DSS

Hammersmith Broadway
2 months ago
Applications closed

Related Jobs

View all jobs

Principal Data Engineer (GCP)

Contract Data Architect/Engineer

Contract Data Architect/Engineer

Contract Data Architect/Engineer

Contract Data Architect/Engineer

Senior Consultant - GCP Data Engineer

Job Title: Data Engineer – GCP/DSS

Department: Enabling Functions

Location: Hybrid, London

Type: Both Contract (Inside IR35) & Permanent available

Salary: Competitive; depends on experience and open to discussion

Purpose of Job

What you will be working on

While our broker platform is the core technology crucial to success – this role will focus on supporting the middle/back-office operations that will lay the foundations for further and sustained success.

We're a multi-disciplined team, bringing together expertise in software and data engineering, full stack development, platform operations, algorithm research, and data science. Our squads focus on delivering high-impact solutions – we favour a highly iterative, analytical approach.

You will be designing and developing complex data processing modules and reporting using Big Query and Tableau. In addition, you will also work closely with the Infrastructure/Platform Team, responsible for architecting and operating the core of the Data Analytics platform.

Principle Accountabilities

Work with both the business teams (finance and actuary initially), data scientists and engineers to design, build, optimise and maintain production grade data pipelines and reporting from an internal Data warehouse solution, based on GCP/Big Query.

Work with finance, actuaries, data scientists and engineers to understand how we can make best use of new internal and external data sources.

Work with our delivery partners at EY/IBM to ensure robustness of design and engineering of the data model/MI and reporting which can support our ambitions for growth and scale.

BAU ownership of data models, reporting and integrations/pipelines.

Create frameworks, infrastructure and systems to manage and govern data assets.

Produce detailed documentation to allow ongoing BAU support and maintenance of data structures, schema, reporting etc.

Work with the broader Engineering community to develop our data and MLOps capability infrastructure.

Ensure data quality, governance, and compliance with internal and external standards.

Monitor and troubleshoot data pipeline issues, ensuring reliability and accuracy

Regulatory Conduct and Rules

  1. Act with integrity

  2. Act with due skill, care and diligence

  3. Be open and co-operative with Lloyd’s, the FCA, the PRA, and other regulators

  4. Pay due regard to the interests of customers and treat them fairly

  5. Observe proper standards of market conduct

    Education, Qualifications, Knowledge, Skills and Experience

  • Experience designing data models and developing industrialised data pipelines.

  • Strong knowledge of database and data lake systems.

  • Hands-on experience in Big Query, dbt, GCP cloud storage.

  • Proficient in Python, SQL and Terraform.

  • Knowledge of Cloud SQL, Airbyte, Dagster.

  • Comfortable with shell scripting with Bash or similar.

  • Experience provisioning new infrastructure in a leading cloud provider, preferably GCP.

  • Proficient with Tableau Cloud for data visualization and reporting.

  • Experience creating DataOps pipelines.

  • Comfortable working in an Agile environment, actively participating in approaches such as Scrum or Kanban

    Desirable Skills

    Experience of streaming data systems and frameworks would be a plus.

    Experience working in regulated industry, especially financial services, would be a plus.

    Experience creating MLOps pipelines is a plus

    The applicant must also demonstrate the following skills and abilities

    Excellent communication skills (both oral and written).

    Pro-active, self-motivated and able to use own initiative.

    Excellent analytical and technical skills.

    Ability to quickly comprehend the functions and capabilities of new technologies.

    Ability to offer balanced opinion regarding existing and future technologies.

    How to Apply

    If you are interested in the Data Engineer – GCP/DSS position, please apply here

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Data Science Tools Do You Need to Know to Get a Data Science Job?

If you’re trying to break into data science — or progress your career — it can feel like you are drowning in names: Python, R, TensorFlow, PyTorch, SQL, Spark, AWS, Scikit-learn, Jupyter, Tableau, Power BI…the list just keeps going. With every job advert listing a different combination of tools, many applicants fall into a trap: they try to learn everything. The result? Long tool lists that sound impressive — but little depth to back them up. Here’s the straight-talk version most hiring managers won’t explicitly tell you: 👉 You don’t need to know every data science tool to get hired. 👉 You need to know the right ones — deeply — and know how to use them to solve real problems. Tools matter, but only in service of outcomes. So how many data science tools do you actually need to know to get a job? For most job seekers, the answer is not “27” — it’s more like 8–12, thoughtfully chosen and well understood. This guide explains what employers really value, which tools are core, which are role-specific, and how to focus your toolbox so your CV and interviews shine.

What Hiring Managers Look for First in Data Science Job Applications (UK Guide)

If you’re applying for data science roles in the UK, it’s crucial to understand what hiring managers focus on before they dive into your full CV. In competitive markets, recruiters and hiring managers often make their first decisions in the first 10–20 seconds of scanning an application — and in data science, there are specific signals they look for first. Data science isn’t just about coding or statistics — it’s about producing insights, shipping models, collaborating with teams, and solving real business problems. This guide helps you understand exactly what hiring managers look for first in data science applications — and how to structure your CV, portfolio and cover letter so you leap to the top of the shortlist.

The Skills Gap in Data Science Jobs: What Universities Aren’t Teaching

Data science has become one of the most visible and sought-after careers in the UK technology market. From financial services and retail to healthcare, media, government and sport, organisations increasingly rely on data scientists to extract insight, guide decisions and build predictive models. Universities have responded quickly. Degrees in data science, analytics and artificial intelligence have expanded rapidly, and many computer science courses now include data-focused pathways. And yet, despite the volume of graduates entering the market, employers across the UK consistently report the same problem: Many data science candidates are not job-ready. Vacancies remain open. Hiring processes drag on. Candidates with impressive academic backgrounds fail interviews or struggle once hired. The issue is not intelligence or effort. It is a persistent skills gap between university education and real-world data science roles. This article explores that gap in depth: what universities teach well, what they often miss, why the gap exists, what employers actually want, and how jobseekers can bridge the divide to build successful careers in data science.