Senior Data Engineer (Trade Surveillance)

TP ICAP
Belfast
5 days ago
Create job alert
Role Overview

Effective Market Abuse Surveillance is highly dependent on data of multiple types and from many different sources. Trade Surveillance requires complete Trade and Pre‑Trade transactional data, Instrument and Client Reference Data and Market Data. Communications Surveillance requires ingestion of messages from multiple electronic platforms. This role will play a critical role in ensuring the reliability, scalability, and compliance of data pipelines that support surveillance systems across communications and trading activities, covering the above structured and unstructured data. The role bridges engineering and operations, enabling robust data ingestion, transformation, and monitoring to meet regulatory and internal compliance requirements. The Data Ops Engineer will collaborate with upstream teams to ensure data completeness, accuracy, and timeliness, and that any data completeness or quality issues are visible. The role will also work on other Surveillance data initiatives such as persisting Surveillance Alerts in the firm’s data lake for analytics purposes.


Job Band Manager, Job Level 6


Role Responsibilities

  • Design, build, maintain and optimise end‑to‑end data pipelines and workflows between the source data points and target destinations, working with the wider Surveillance Technology team to prioritise automation, scalability and strategy at the heart of the design.
  • Implement automated data completeness and quality checks, validation rules, and reconciliation processes to ensure accuracy, completeness, and timeliness of the data ingested and to make visible any data that is not processed.
  • Identify Critical Data Elements and implement failover and recovery strategies for the respective Data Flows.
  • Build AWS infrastructure using Terraform or CDK.
  • Write unit, integration, and infrastructure tests.
  • Monitor, investigate and resolve data anomalies through collaboration with Business Analysts, Developers, and Testers across functions and verticals.
  • Implement data management and governance frameworks to ensure data is ingested and loaded per the requirements of the consuming platform; Scila for Trade Surveillance, and Global Relay for Communications Surveillance.
  • Partner with the Data Strategy and Data Infrastructure team to ensure Data Lineage, auditability and retention policies are enforced across all necessary pipelines.
  • Ensure that data consumed and processed is compliance with regulatory, legal, and security protocols.
  • Work closely with surveillance analysts, compliance officers, and engineering teams to translate business rules into technical specifications.
  • Partner closely with stakeholders and subject matter experts such as the Cloud Infrastructure team to optimise performance and costs.
  • Stay updated on industry trends and emerging tech to ensure continuous improvement.

Experience / Competences
Essential Criteria

  • Strong experience ETL/ELT data pipeline builds from design, to implementation, to maintenance in relation to financial market messaging platforms, and trade & order systems.
  • Solid understanding of CI/CD pipelines, ideally with a background in software engineering, product management or data analytics.
  • Experience with EKS, Lambda, EventBridge, Step Functions, S3, DynamoDB, AWS Glue, Snowflake, Terraform and Transfer Family.
  • Strong proficiency in Python or Java, SQL, and data pipeline frameworks (e.g., Airflow, dbt, Spark), with solid experience with the AWS ecosystem.
  • Proven expertise with data governance frameworks and compliance regulations in financial services.
  • Knowledge of streaming technologies (Kafka, Kinesis) and API integrations, and hands‑on experience with monitoring tools (e.g. Grafana) and observability practices.
  • Excellent problem‑solving skills and ability to work in a fast‑paced environment.
  • Bachelor’s degree in Computer Science, Data Science, Engineering, or related field.
  • Previous experience in Data Ops and Data Engineering.
  • Strong communication and collaboration skills to engage with technical and non‑technical stakeholders.
  • Strong experience with Agile software delivery.

Non Essential

  • Experience with market data ingestion, metadata extraction, and event‑driven architectures.
  • Proficient with Terraform or CDK (infrastructure‑as‑code).
  • Experience in Business Communications Technology e.g. Bloomberg, ICE, Symphony, Teams Chat, etc.
  • Familiarity with security best practices, IAM, and VPN configuration.
  • Experience with regulatory compliance and data security in financial services.
  • Knowledge of financial markets and trading platforms.
  • Experience with GitLab, Qliksense & Alation.
  • Certifications in DataOps, cloud platforms, or related areas.

Not The Perfect Fit?

Concerned that you may not meet the criteria precisely? At TP ICAP, we wholeheartedly believe in fostering inclusivity and cultivating a work environment where everyone can flourish, regardless of your personal or professional background. If you are enthusiastic about this role but find that your experience doesn't align perfectly with every aspect of the job description, we strongly encourage you to apply. You may be the ideal candidate for this position or another opportunity within our organisation. Our dedicated Talent Acquisition team is here to assist you in recognising how your unique skills and abilities can be a valuable contribution. Don't hesitate to take the leap and explore the possibilities. Your potential is what truly matters to us.


Company Statement

We know that the best innovation happens when diverse people with different perspectives and skills work together in an inclusive atmosphere. That's why we're building a culture where everyone plays a part in making people feel welcome, ready and willing to contribute. TP ICAP Accord – our Employee Network – is a central part of this. As well as representing specific groups, TP ICAP Accord helps increase awareness, collaboration, shares best practice, and holds our firm to account for driving continuous cultural improvement.


Location

UK – City Quays – Belfast


#J-18808-Ljbffr

Related Jobs

View all jobs

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Senior Data Engineer

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Data Science Tools Do You Need to Know to Get a Data Science Job?

If you’re trying to break into data science — or progress your career — it can feel like you are drowning in names: Python, R, TensorFlow, PyTorch, SQL, Spark, AWS, Scikit-learn, Jupyter, Tableau, Power BI…the list just keeps going. With every job advert listing a different combination of tools, many applicants fall into a trap: they try to learn everything. The result? Long tool lists that sound impressive — but little depth to back them up. Here’s the straight-talk version most hiring managers won’t explicitly tell you: 👉 You don’t need to know every data science tool to get hired. 👉 You need to know the right ones — deeply — and know how to use them to solve real problems. Tools matter, but only in service of outcomes. So how many data science tools do you actually need to know to get a job? For most job seekers, the answer is not “27” — it’s more like 8–12, thoughtfully chosen and well understood. This guide explains what employers really value, which tools are core, which are role-specific, and how to focus your toolbox so your CV and interviews shine.

What Hiring Managers Look for First in Data Science Job Applications (UK Guide)

If you’re applying for data science roles in the UK, it’s crucial to understand what hiring managers focus on before they dive into your full CV. In competitive markets, recruiters and hiring managers often make their first decisions in the first 10–20 seconds of scanning an application — and in data science, there are specific signals they look for first. Data science isn’t just about coding or statistics — it’s about producing insights, shipping models, collaborating with teams, and solving real business problems. This guide helps you understand exactly what hiring managers look for first in data science applications — and how to structure your CV, portfolio and cover letter so you leap to the top of the shortlist.

The Skills Gap in Data Science Jobs: What Universities Aren’t Teaching

Data science has become one of the most visible and sought-after careers in the UK technology market. From financial services and retail to healthcare, media, government and sport, organisations increasingly rely on data scientists to extract insight, guide decisions and build predictive models. Universities have responded quickly. Degrees in data science, analytics and artificial intelligence have expanded rapidly, and many computer science courses now include data-focused pathways. And yet, despite the volume of graduates entering the market, employers across the UK consistently report the same problem: Many data science candidates are not job-ready. Vacancies remain open. Hiring processes drag on. Candidates with impressive academic backgrounds fail interviews or struggle once hired. The issue is not intelligence or effort. It is a persistent skills gap between university education and real-world data science roles. This article explores that gap in depth: what universities teach well, what they often miss, why the gap exists, what employers actually want, and how jobseekers can bridge the divide to build successful careers in data science.