Staff Data Engineer - Data Quality & Governance

Depop
City of London
1 month ago
Applications closed

Related Jobs

View all jobs

Staff Data Engineer

Staff Data Engineer

Staff Data Engineer

Senior Data Engineer - Azure

Senior Data Engineer (AWS, Airflow, Python)

Senior/Lead Data Engineer

Company Description

Depop is the community-powered circular fashion marketplace where anyone can buy, sell and discover desirable secondhand fashion. With a community of over 35 million users, Depop is on a mission to make fashion circular, redefining fashion consumption. Founded in 2011, the company is headquartered in London, with offices in New York and Manchester, and in 2021 became a wholly-owned subsidiary of Etsy. Find out more at www.depop.com.


Our mission is to make fashion circular and to create an inclusive environment where everyone is welcome, no matter who they are or where they’re from. Just as our platform connects people globally, we believe our workplace should reflect the diversity of the communities we serve. We thrive on the power of different perspectives and experiences, knowing they drive innovation and bring us closer to our users. We're proud to be an equal opportunity employer, providing employment opportunities without regard to age, ethnicity, religion or belief, gender identity, sex, sexual orientation, disability, pregnancy or maternity, marriage and civil partnership, or any other protected status. We're continuously evolving our recruitment processes to ensure fairness and are open to accommodating any needs you might have.


If, due to a disability, you need adjustments to complete the application, please let us know by sending an email with your name, the role to which you would like to apply, and the type of support you need to complete the application to .


Role

We're building a Data Quality, Observability & Governance Team to improve the reliability, trust, and compliance of Depop's data ecosystem.


As a Staff Data Engineer in this team, you'll lead the design and implementation of frameworks, tools, and processes that strengthen our data foundations - ensuring our data is accurate, observable, and compliant.


Your mission will be to reduce the mean time to detection and resolution of data incidents, by establishing data contracts between producers and consumers, developing robust data observability systems, and embedding governance and GDPR compliance principles across the data lifecycle.


You'll collaborate with product engineering, data platform, analytics, and legal teams to build confidence in data as a product - one that's reliable, auditable, and actionable.


Responsibilities

  • Define and execute the vision for Depop's data quality, observability, and governance frameworks.
  • Establish data contracts between producers and consumers to ensure schema integrity and data reliability.
  • Develop and maintain systems that detect, alert, and resolve data quality issues with minimal latency.
  • Build automation and tooling to reduce MTTD (Mean Time to Detection) and MTTR (Mean Time to Resolution) for data incidents.
  • Partner with data platform engineers to integrate observability at every layer - ingestion, transformation, and consumption.
  • Lead GDPR and privacy-by-design initiatives, ensuring compliance and traceability across all datasets.
  • Define standards for metadata management, lineage tracking, and access control.
  • Collaborate with analytics and product teams to ensure data definitions and quality metrics are consistent across domains.
  • Mentor engineers and analysts, fostering a culture of data stewardship and accountability.
  • Continuously improve data governance maturity through automation, documentation, and measurable quality KPIs.

Qualifications

  • Proven experience as a Staff Data Engineer or in an equivalent technical leadership role in data quality, observability, or governance.
  • Deep knowledge of data observability frameworks (Monte Carlo, Soda, or equivalent) and data validation tools (Great Expectations, DBT tests, etc.).
  • Deep understanding of data-as-a-product principles and experience applying them to improve data reliability and ownership.
  • Experience designing and enforcing data contracts and quality SLAs in distributed data ecosystems.
  • Proficiency in Python, Java, or Scala, and experience building pipelines with Databricks, Spark, or Kafka.
  • Strong understanding of data governance principles, privacy regulations (GDPR, CCPA), and secure data handling practices.
  • Familiarity with metadata management and data catalog tools (e.g. DataHub, Collibra, etc.).
  • Demonstrated success improving data reliability and observability in large-scale data platforms.
  • Excellent communication and stakeholder management skills; you can bridge technical depth with operational impact.

Bonus Points

  • Experience implementing automated compliance monitoring or policy-as-code systems.
  • Familiarity with real-time anomaly detection for data pipelines.
  • Experience contributing to or leading cross-functional data reliability initiatives.
  • Prior experience in consumer or marketplace platforms.
  • Passion for data as a product - building reliable, observable, and compliant data systems that teams love to use.

Additional Information
Health + Mental Wellbeing

  • PMI and cash plan healthcare access with Bupa
  • Subsidised counselling and coaching with Self Space
  • Cycle to Work scheme with options from Evans or the Green Commute Initiative
  • Employee Assistance Programme (EAP) for 24/7 confidential support
  • Mental Health First Aiders across the business for support and signposting

Work/Life Balance

  • 25 days annual leave with option to carry over up to 5 days
  • 1 company-wide day off per quarter
  • Impact hours: Up to 2 days additional paid leave per year for volunteering
  • Fully paid 4 week sabbatical after completion of 5 years of consecutive service with Depop, to give you a chance to recharge or do something you love.
  • Flexible Working: MyMode hybrid-working model with Flex, Office Based, and Remote options *role dependant
  • All offices are dog-friendly
  • Ability to work abroad for 4 weeks per year in UK tax treaty countries

Family Life

  • 18 weeks of paid parental leave for full-time regular employees
  • IVF leave, shared parental leave, and paid emergency parent/carer leave

Learn + Grow

  • Budgets for conferences, learning subscriptions, and more
  • Mentorship and programmes to upskill employees

Your Future

  • Life Insurance (financial compensation of 3x your salary)
  • Pension matching up to 6% of qualifying earnings

Depop Extras

  • Employees enjoy free shipping on their Depop sales within the UK.
  • Special milestones are celebrated with gifts and rewards!


#J-18808-Ljbffr

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Data Science Tools Do You Need to Know to Get a Data Science Job?

If you’re trying to break into data science — or progress your career — it can feel like you are drowning in names: Python, R, TensorFlow, PyTorch, SQL, Spark, AWS, Scikit-learn, Jupyter, Tableau, Power BI…the list just keeps going. With every job advert listing a different combination of tools, many applicants fall into a trap: they try to learn everything. The result? Long tool lists that sound impressive — but little depth to back them up. Here’s the straight-talk version most hiring managers won’t explicitly tell you: 👉 You don’t need to know every data science tool to get hired. 👉 You need to know the right ones — deeply — and know how to use them to solve real problems. Tools matter, but only in service of outcomes. So how many data science tools do you actually need to know to get a job? For most job seekers, the answer is not “27” — it’s more like 8–12, thoughtfully chosen and well understood. This guide explains what employers really value, which tools are core, which are role-specific, and how to focus your toolbox so your CV and interviews shine.

What Hiring Managers Look for First in Data Science Job Applications (UK Guide)

If you’re applying for data science roles in the UK, it’s crucial to understand what hiring managers focus on before they dive into your full CV. In competitive markets, recruiters and hiring managers often make their first decisions in the first 10–20 seconds of scanning an application — and in data science, there are specific signals they look for first. Data science isn’t just about coding or statistics — it’s about producing insights, shipping models, collaborating with teams, and solving real business problems. This guide helps you understand exactly what hiring managers look for first in data science applications — and how to structure your CV, portfolio and cover letter so you leap to the top of the shortlist.

The Skills Gap in Data Science Jobs: What Universities Aren’t Teaching

Data science has become one of the most visible and sought-after careers in the UK technology market. From financial services and retail to healthcare, media, government and sport, organisations increasingly rely on data scientists to extract insight, guide decisions and build predictive models. Universities have responded quickly. Degrees in data science, analytics and artificial intelligence have expanded rapidly, and many computer science courses now include data-focused pathways. And yet, despite the volume of graduates entering the market, employers across the UK consistently report the same problem: Many data science candidates are not job-ready. Vacancies remain open. Hiring processes drag on. Candidates with impressive academic backgrounds fail interviews or struggle once hired. The issue is not intelligence or effort. It is a persistent skills gap between university education and real-world data science roles. This article explores that gap in depth: what universities teach well, what they often miss, why the gap exists, what employers actually want, and how jobseekers can bridge the divide to build successful careers in data science.