Lead Product Data Engineer

Inovia Bio
London
3 months ago
Applications closed

Related Jobs

View all jobs

Lead Data Engineer, Data Reliability

Head of Data Engineering (Basé à London)

Head of Data Engineering

Lead Data Engineer

Lead Data Engineer

Lead Data Engineer

Job Title:Lead Product Data Engineer

Location:London, UK (Hybrid)

Salary:£95,000 - £105,000/year + options + private medical insurance + unlimited holiday

Company:Inovia Bio

This role is for candidates only, not recruiters.


About Us:


At Inovia Bio we believe that drug development is in dire need of disruptive thinking.

The current ecosystem supporting biopharmaceutical companies is typified by slow, archaic incumbents that extract as much value as possible from drug developers and their companies while often delivering utter garbage. We’re changing that!

From the way data is used to make early decisions through to how development plans can be accelerated and de-risked, there are significant opportunities to get drugs to patients faster. Your role will be central to our mission of revolutionising drug development.


Who are we?


Inovia Bio is a TechBio Startup based in London. We partner with innovative biopharmaceutical companies by providing both our proprietary technology platforms and expertise to deliver unprecedented impact to their programs. We are the company drug developers dream of working with and the company CROs dread.


An Atypical Role:


The Lead Product Data Engineer at Inovia operates differently –We believe that drug development can only be solved by the merging of medicine and technology…and so will you!


We are seeking a seasonedLead Product Data Engineerwith a strong product-oriented mindset, that fully embraces collaboration with medical colleagues. You will feel comfortable going into a new industry and learning is a key part of your nature.


You will take ownership of end-to-end data processes and embrace collaboration. You will work closely with cross-functional teams to design, build, and maintain robust data infrastructure on the Google Cloud Platform (GCP), enabling data-driven decision-making and facilitating our mission to accelerate drug development.


Key Responsibilities:


  • Collaboration: Work with data epidemiologists, product managers, and stakeholders to define requirements, iterate on solutions, and ensure alignment with business objectives.
  • Design & Development: Develop, test, and deploy scalable data pipelines using Python, Flask, Apache Spark/Beam, and GCP Dataflow to meet the needs of our customers while collaborating with the medical, biostatistics and epidemiology teams.
  • Data Orchestration & Automation: Utilize Airflow for workflow orchestration and Cloud Functions for seamless integration with GCP services, automating critical processes.
  • Real-Time Data Processing: Implement data streaming solutions with Pub/Sub or Kafka, ensuring low-latency processing and real-time analytics capabilities.
  • Data Warehouse Optimisation & Modelling: Set up and optimize BigQuery, SQL, DBT models, and VertexAI workflows for advanced data transformations and machine learning model serving.
  • Data Visualization: Collaborate with stakeholders to define requirements and build interactive dashboards on Looker, translating complex datasets into actionable insights.
  • Infrastructure as Code: Use Terraform to define and manage infrastructure in a reproducible manner, ensuring a scalable, secure, and consistent data platform.
  • CI/CD Pipelines: Implement and monitor CI/CD pipelines to support continuous integration and automated deployment of data workflows.
  • Implement and Optimise ML Models: Having a good mathematical understanding of LLMs and general machine learning models.

 

Key Requirements:


·      Collaboration: Work with the medical team, product managers, and stakeholders to define requirements, iterate on solutions, and ensure alignment with business objectives.

·      Product-Oriented Mindset: Ability to prioritize customer needs and translate them into actionable data solutions.

·      You believe in“strong opinions loosely held”: No idea or approach is sacred – you’re comfortable speaking up and challenging ideas and comfortable receiving feedback.

·      Technical Proficiency: Extensive experience with Python, Flask, Airflow, GCP Dataflow, Apache Spark/Beam, Cloud Functions, Pub/Sub(or Kafka), Vertex AI, CI/CD, Linux, BiqQuery, SQL, DBT, Looker, Git and GitFlow.

·      Google Cloud Platform Knowledge: Deep understanding of GCP services and architecture best practices and use of Terraform.

·      You have worked in an Agile environment

·      3+ years of experience

·      Eligibility to work in the UK

·      All successful applicants will need to pass a DBS check.



Good-to-Haves:


  • ML-Ops Experience: Familiarity with ML model deployment and monitoring, especially within a GCP/VertexAI environment.
  • Full-Stack Development Skills: Experience with web development frameworks and integrating front-end and back-end services.
  • Web Crawling: Hands-on experience with web scraping and data extraction techniques.
  • Looker Dashboard Expertise: Proven experience in building and optimizing Looker dashboards for data storytelling.
  • 3+ years of experience


What we offer:

  • Competitive salary in the range of £95,000 to £105,000/year.
  • Equity options to share in our success.
  • Private medical insurance.
  • Unlimited holiday policy.
  • An opportunity to work on impactful data projects in a dynamic and innovative environment.
  • Hybrid work environment.
  • Rapid progression opportunities
  • Zero political work culture.
  • Learning budget

Get the latest insights and jobs direct. Sign up for our newsletter.

By subscribing you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

Portfolio Projects That Get You Hired for Data Science Jobs (With Real GitHub Examples)

Data science is at the forefront of innovation, enabling organisations to turn vast amounts of data into actionable insights. Whether it’s building predictive models, performing exploratory analyses, or designing end-to-end machine learning solutions, data scientists are in high demand across every sector. But how can you stand out in a crowded job market? Alongside a solid CV, a well-curated data science portfolio often makes the difference between getting an interview and getting overlooked. In this comprehensive guide, we’ll explore: Why a data science portfolio is essential for job seekers. Selecting projects that align with your target data science roles. Real GitHub examples showcasing best practices. Actionable project ideas you can build right now. Best ways to present your projects and ensure recruiters can find them easily. By the end, you’ll be equipped to craft a compelling portfolio that proves your skills in a tangible way. And when you’re ready for your next career move, remember to upload your CV on DataScience-Jobs.co.uk so that your newly showcased work can be discovered by employers looking for exactly what you have to offer.

Data Science Job Interview Warm‑Up: 30 Real Coding & System‑Design Questions

Data science has become one of the most sought‑after fields in technology, leveraging mathematics, statistics, machine learning, and programming to derive valuable insights from data. Organisations across every sector—finance, healthcare, retail, government—rely on data scientists to build predictive models, understand patterns, and shape strategy with data‑driven decisions. If you’re gearing up for a data science interview, expect a well‑rounded evaluation. Beyond statistics and algorithms, many roles also require data wrangling, visualisation, software engineering, and communication skills. Interviewers want to see if you can slice and dice messy datasets, design experiments, and scale ML models to production. In this guide, we’ll explore 30 real coding & system‑design questions commonly posed in data science interviews. You’ll find challenges ranging from algorithmic coding and statistical puzzle‑solving to the architectural side of building data science platforms in real‑world settings. By practising with these questions, you’ll gain the confidence and clarity needed to stand out among competitive candidates. And if you’re actively seeking data science opportunities in the UK, be sure to visit www.datascience-jobs.co.uk. It’s a comprehensive hub featuring junior, mid‑level, and senior data science vacancies—spanning start‑ups to FTSE 100 companies. Let’s dive into what you need to know.

Negotiating Your Data Science Job Offer: Equity, Bonuses & Perks Explained

Data science has rapidly evolved from a niche specialty to a cornerstone of strategic decision-making in virtually every industry—from finance and healthcare to retail, entertainment, and AI research. As a mid‑senior data scientist, you’re not just running predictive models or generating dashboards; you’re shaping business strategy, product innovation, and customer experiences. This level of influence is why employers are increasingly offering compensation packages that go beyond a baseline salary. Yet, many professionals still tend to focus almost exclusively on base pay when negotiating a new role. This can be a costly oversight. Companies vying for data science talent—especially in the UK, where demand often outstrips supply—routinely offer equity, bonuses, flexible work options, and professional development funds in addition to salary. Recognising these opportunities and effectively negotiating them can have a substantial impact on your total earnings and long-term career satisfaction. This guide explores every facet of negotiating a data science job offer—from understanding equity structures and bonus schemes to weighing crucial perks like remote work and ongoing skill development. By the end, you’ll be well-equipped to secure a holistic package aligned with your market value, your life goals, and the tremendous impact you bring to any organisation.